var/home/core/zuul-output/0000755000175000017500000000000015111774112014525 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112000364015461 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004272222715112000355017675 0ustar rootrootNov 27 07:08:35 crc systemd[1]: Starting Kubernetes Kubelet... Nov 27 07:08:35 crc restorecon[4680]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 07:08:35 crc restorecon[4680]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 27 07:08:36 crc kubenswrapper[4706]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 07:08:36 crc kubenswrapper[4706]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 27 07:08:36 crc kubenswrapper[4706]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 07:08:36 crc kubenswrapper[4706]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 07:08:36 crc kubenswrapper[4706]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 27 07:08:36 crc kubenswrapper[4706]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.510915 4706 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514395 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514417 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514424 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514430 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514436 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514442 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514448 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514454 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514459 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514469 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514477 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514484 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514491 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514498 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514506 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514514 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514521 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514527 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514533 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514538 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514544 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514555 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514561 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514568 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514575 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514580 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514585 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514591 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514596 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514601 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514607 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514613 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514618 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514623 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514629 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514635 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514640 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514646 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514652 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514657 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514662 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514668 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514673 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514678 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514685 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514692 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514699 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514704 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514711 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514717 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514723 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514729 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514735 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514741 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514747 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514753 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514759 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514764 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514770 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514775 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514780 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514785 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514791 4706 feature_gate.go:330] unrecognized feature gate: Example Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514796 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514801 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514806 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514812 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514817 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514823 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514829 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.514836 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.514968 4706 flags.go:64] FLAG: --address="0.0.0.0" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.514979 4706 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.514994 4706 flags.go:64] FLAG: --anonymous-auth="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515002 4706 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515009 4706 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515015 4706 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515022 4706 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515029 4706 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515036 4706 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515042 4706 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515050 4706 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515057 4706 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515063 4706 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515069 4706 flags.go:64] FLAG: --cgroup-root="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515075 4706 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515081 4706 flags.go:64] FLAG: --client-ca-file="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515086 4706 flags.go:64] FLAG: --cloud-config="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515092 4706 flags.go:64] FLAG: --cloud-provider="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515098 4706 flags.go:64] FLAG: --cluster-dns="[]" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515105 4706 flags.go:64] FLAG: --cluster-domain="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515111 4706 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515117 4706 flags.go:64] FLAG: --config-dir="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515124 4706 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515130 4706 flags.go:64] FLAG: --container-log-max-files="5" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515138 4706 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515144 4706 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515150 4706 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515156 4706 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515162 4706 flags.go:64] FLAG: --contention-profiling="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515168 4706 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515174 4706 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515180 4706 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515186 4706 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515193 4706 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515199 4706 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515205 4706 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515210 4706 flags.go:64] FLAG: --enable-load-reader="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515231 4706 flags.go:64] FLAG: --enable-server="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515238 4706 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515246 4706 flags.go:64] FLAG: --event-burst="100" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515252 4706 flags.go:64] FLAG: --event-qps="50" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515258 4706 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515264 4706 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515270 4706 flags.go:64] FLAG: --eviction-hard="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515276 4706 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515282 4706 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515299 4706 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515306 4706 flags.go:64] FLAG: --eviction-soft="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515312 4706 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515318 4706 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515324 4706 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515353 4706 flags.go:64] FLAG: --experimental-mounter-path="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515359 4706 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515365 4706 flags.go:64] FLAG: --fail-swap-on="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515371 4706 flags.go:64] FLAG: --feature-gates="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515380 4706 flags.go:64] FLAG: --file-check-frequency="20s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515386 4706 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515393 4706 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515400 4706 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515407 4706 flags.go:64] FLAG: --healthz-port="10248" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515414 4706 flags.go:64] FLAG: --help="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515420 4706 flags.go:64] FLAG: --hostname-override="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515426 4706 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515433 4706 flags.go:64] FLAG: --http-check-frequency="20s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515440 4706 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515447 4706 flags.go:64] FLAG: --image-credential-provider-config="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515453 4706 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515459 4706 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515466 4706 flags.go:64] FLAG: --image-service-endpoint="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515471 4706 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515477 4706 flags.go:64] FLAG: --kube-api-burst="100" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515484 4706 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515490 4706 flags.go:64] FLAG: --kube-api-qps="50" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515496 4706 flags.go:64] FLAG: --kube-reserved="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515502 4706 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515507 4706 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515513 4706 flags.go:64] FLAG: --kubelet-cgroups="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515519 4706 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515524 4706 flags.go:64] FLAG: --lock-file="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515530 4706 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515536 4706 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515542 4706 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515551 4706 flags.go:64] FLAG: --log-json-split-stream="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515556 4706 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515562 4706 flags.go:64] FLAG: --log-text-split-stream="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515569 4706 flags.go:64] FLAG: --logging-format="text" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515575 4706 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515581 4706 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515587 4706 flags.go:64] FLAG: --manifest-url="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515592 4706 flags.go:64] FLAG: --manifest-url-header="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515604 4706 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515609 4706 flags.go:64] FLAG: --max-open-files="1000000" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515616 4706 flags.go:64] FLAG: --max-pods="110" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515622 4706 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515628 4706 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515634 4706 flags.go:64] FLAG: --memory-manager-policy="None" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515640 4706 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515646 4706 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515651 4706 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515657 4706 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515671 4706 flags.go:64] FLAG: --node-status-max-images="50" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515677 4706 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515684 4706 flags.go:64] FLAG: --oom-score-adj="-999" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515690 4706 flags.go:64] FLAG: --pod-cidr="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515698 4706 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515707 4706 flags.go:64] FLAG: --pod-manifest-path="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515713 4706 flags.go:64] FLAG: --pod-max-pids="-1" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515719 4706 flags.go:64] FLAG: --pods-per-core="0" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515725 4706 flags.go:64] FLAG: --port="10250" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515730 4706 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515736 4706 flags.go:64] FLAG: --provider-id="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515742 4706 flags.go:64] FLAG: --qos-reserved="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515748 4706 flags.go:64] FLAG: --read-only-port="10255" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515754 4706 flags.go:64] FLAG: --register-node="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515759 4706 flags.go:64] FLAG: --register-schedulable="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515765 4706 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515774 4706 flags.go:64] FLAG: --registry-burst="10" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515780 4706 flags.go:64] FLAG: --registry-qps="5" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515786 4706 flags.go:64] FLAG: --reserved-cpus="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515791 4706 flags.go:64] FLAG: --reserved-memory="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515798 4706 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515804 4706 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515810 4706 flags.go:64] FLAG: --rotate-certificates="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515816 4706 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515822 4706 flags.go:64] FLAG: --runonce="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515827 4706 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515833 4706 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515839 4706 flags.go:64] FLAG: --seccomp-default="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515845 4706 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515851 4706 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515857 4706 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515862 4706 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515868 4706 flags.go:64] FLAG: --storage-driver-password="root" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515874 4706 flags.go:64] FLAG: --storage-driver-secure="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515880 4706 flags.go:64] FLAG: --storage-driver-table="stats" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515886 4706 flags.go:64] FLAG: --storage-driver-user="root" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515893 4706 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515899 4706 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515905 4706 flags.go:64] FLAG: --system-cgroups="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515910 4706 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515919 4706 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515924 4706 flags.go:64] FLAG: --tls-cert-file="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515930 4706 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515936 4706 flags.go:64] FLAG: --tls-min-version="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515942 4706 flags.go:64] FLAG: --tls-private-key-file="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515951 4706 flags.go:64] FLAG: --topology-manager-policy="none" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515957 4706 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515962 4706 flags.go:64] FLAG: --topology-manager-scope="container" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515967 4706 flags.go:64] FLAG: --v="2" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515975 4706 flags.go:64] FLAG: --version="false" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515982 4706 flags.go:64] FLAG: --vmodule="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515989 4706 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.515995 4706 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516128 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516135 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516141 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516147 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516152 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516157 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516162 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516167 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516172 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516176 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516181 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516186 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516191 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516197 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516203 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516208 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516213 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516236 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516241 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516246 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516251 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516256 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516261 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516266 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516274 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516279 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516284 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516289 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516294 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516298 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516303 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516308 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516313 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516318 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516324 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516330 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516335 4706 feature_gate.go:330] unrecognized feature gate: Example Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516341 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516346 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516351 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516356 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516361 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516366 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516371 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516375 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516380 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516385 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516389 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516394 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516399 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516404 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516409 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516413 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516418 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516425 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516431 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516438 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516444 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516449 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516456 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516461 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516468 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516474 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516479 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516485 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516490 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516494 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516499 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516504 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516510 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.516516 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.516524 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.532307 4706 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.532371 4706 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532531 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532548 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532558 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532568 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532577 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532589 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532601 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532611 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532620 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532629 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532638 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532646 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532656 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532664 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532673 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532682 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532690 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532699 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532708 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532717 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532725 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532734 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532742 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532751 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532759 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532769 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532777 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532786 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532794 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532802 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532815 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532826 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532837 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532846 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532869 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532878 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532889 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532904 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532914 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532923 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532935 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532944 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532954 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532963 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532972 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532982 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.532991 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533000 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533008 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533016 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533025 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533034 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533045 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533055 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533064 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533072 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533081 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533090 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533098 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533107 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533116 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533126 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533138 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533148 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533158 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533166 4706 feature_gate.go:330] unrecognized feature gate: Example Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533176 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533184 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533193 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533201 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533245 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.533262 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533585 4706 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533609 4706 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533618 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533627 4706 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533636 4706 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533645 4706 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533653 4706 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533662 4706 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533670 4706 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533691 4706 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533700 4706 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533709 4706 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533717 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533726 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533734 4706 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533743 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533751 4706 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533760 4706 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533769 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533778 4706 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533786 4706 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533795 4706 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533803 4706 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533811 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533820 4706 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533829 4706 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533838 4706 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533846 4706 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533855 4706 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533863 4706 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533871 4706 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533880 4706 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533888 4706 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533897 4706 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533918 4706 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533927 4706 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533936 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533945 4706 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533954 4706 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533962 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533970 4706 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533979 4706 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533987 4706 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.533996 4706 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534004 4706 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534015 4706 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534025 4706 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534034 4706 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534043 4706 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534051 4706 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534060 4706 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534068 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534077 4706 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534085 4706 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534093 4706 feature_gate.go:330] unrecognized feature gate: Example Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534102 4706 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534113 4706 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534125 4706 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534136 4706 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534145 4706 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534154 4706 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534166 4706 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534177 4706 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534186 4706 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534196 4706 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534208 4706 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534217 4706 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534254 4706 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534264 4706 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534273 4706 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.534294 4706 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.534308 4706 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.534719 4706 server.go:940] "Client rotation is on, will bootstrap in background" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.541850 4706 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.542001 4706 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.544727 4706 server.go:997] "Starting client certificate rotation" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.544779 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.545118 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-31 14:50:41.106724161 +0000 UTC Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.545411 4706 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 823h42m4.561322252s for next certificate rotation Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.570452 4706 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.577563 4706 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.593612 4706 log.go:25] "Validated CRI v1 runtime API" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.626706 4706 log.go:25] "Validated CRI v1 image API" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.631532 4706 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.639195 4706 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-27-07-04-46-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.639300 4706 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.671506 4706 manager.go:217] Machine: {Timestamp:2025-11-27 07:08:36.666134831 +0000 UTC m=+0.555725721 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:70cebc65-54fc-48ee-bf03-1e5c2a992650 BootID:70ab140b-da2b-4616-bf57-a58441a0c7b5 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:fb:75:a8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:fb:75:a8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:74:c9:dc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a2:cd:e2 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d8:e4:79 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:15:d4:f7 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:e6:57:35:20:fe:31 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8a:e4:13:a0:89:22 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.671901 4706 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.672093 4706 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.672790 4706 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.673174 4706 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.673274 4706 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.673680 4706 topology_manager.go:138] "Creating topology manager with none policy" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.673702 4706 container_manager_linux.go:303] "Creating device plugin manager" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.674465 4706 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.674544 4706 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.676361 4706 state_mem.go:36] "Initialized new in-memory state store" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.676564 4706 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.679784 4706 kubelet.go:418] "Attempting to sync node with API server" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.679826 4706 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.679875 4706 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.679900 4706 kubelet.go:324] "Adding apiserver pod source" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.679922 4706 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.687274 4706 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.690060 4706 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.691401 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.691532 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.691472 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.691677 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.692506 4706 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694515 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694591 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694620 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694640 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694670 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694690 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694713 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694745 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694769 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694789 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694830 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.694846 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.695424 4706 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.696354 4706 server.go:1280] "Started kubelet" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.696628 4706 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.696721 4706 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.697064 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.697559 4706 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 27 07:08:36 crc systemd[1]: Started Kubernetes Kubelet. Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.700331 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.700388 4706 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.700456 4706 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.700480 4706 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.700467 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 03:51:42.603503464 +0000 UTC Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.700599 4706 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.700594 4706 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.701595 4706 server.go:460] "Adding debug handlers to kubelet server" Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.702340 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.702445 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.702866 4706 factory.go:55] Registering systemd factory Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.702912 4706 factory.go:221] Registration of the systemd container factory successfully Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.703727 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.153:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187bcb65b2b91a28 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 07:08:36.69629188 +0000 UTC m=+0.585882730,LastTimestamp:2025-11-27 07:08:36.69629188 +0000 UTC m=+0.585882730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.706497 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="200ms" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.707016 4706 factory.go:153] Registering CRI-O factory Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.707073 4706 factory.go:221] Registration of the crio container factory successfully Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.707384 4706 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.707415 4706 factory.go:103] Registering Raw factory Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.707442 4706 manager.go:1196] Started watching for new ooms in manager Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.708341 4706 manager.go:319] Starting recovery of all containers Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.722093 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723644 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723742 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723770 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723794 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723817 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723841 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723865 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723894 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723916 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723938 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.723962 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724057 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724087 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724110 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724133 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724156 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724182 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724203 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724290 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724343 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724375 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724400 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724421 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724442 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724475 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724506 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724531 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724556 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724577 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724636 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724658 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724680 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724700 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724720 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724742 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724764 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724784 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724808 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724831 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724852 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724875 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724906 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724939 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724969 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.724998 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725028 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725091 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725124 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725152 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725181 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725209 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725286 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725322 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725354 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725385 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725418 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725590 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725625 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725654 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725716 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725749 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725779 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725811 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725840 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725871 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725900 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725933 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.725971 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726001 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726033 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726062 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726090 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726123 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726199 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726269 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726303 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726334 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726369 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726397 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726429 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726463 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726490 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726522 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726558 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726588 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726622 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726650 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726683 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726897 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726933 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726963 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.726997 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727065 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727099 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727130 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727158 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727191 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727256 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727289 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727324 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727352 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727384 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727415 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727457 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727493 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727528 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727570 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727616 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727650 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727679 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727712 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727747 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727784 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727820 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727848 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727877 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727903 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727932 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727959 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.727986 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728011 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728037 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728065 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728091 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728116 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728141 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728166 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728189 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728213 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728288 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728316 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728351 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728378 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728409 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728436 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728466 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728494 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728521 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728553 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728580 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728604 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728632 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728659 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728687 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728758 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728784 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728811 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728839 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728865 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728891 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728916 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728949 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.728974 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729003 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729029 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729056 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729083 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729111 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729139 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729165 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729192 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729261 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729295 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729319 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729340 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729364 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729385 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729405 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729425 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729448 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729471 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729492 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729512 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.729537 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.731854 4706 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.731917 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.731943 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.731967 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.731990 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732020 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732042 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732065 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732090 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732112 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732134 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732155 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732178 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732202 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732262 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732294 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732316 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732338 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732360 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732381 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732404 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732433 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732457 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732484 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732509 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732532 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732555 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732577 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732598 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732623 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732646 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732671 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732694 4706 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732715 4706 reconstruct.go:97] "Volume reconstruction finished" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.732730 4706 reconciler.go:26] "Reconciler: start to sync state" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.744075 4706 manager.go:324] Recovery completed Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.757585 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.759679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.759749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.759773 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.763864 4706 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.763898 4706 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.763934 4706 state_mem.go:36] "Initialized new in-memory state store" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.772038 4706 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.775486 4706 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.775525 4706 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.775549 4706 kubelet.go:2335] "Starting kubelet main sync loop" Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.775596 4706 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 27 07:08:36 crc kubenswrapper[4706]: W1127 07:08:36.779753 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.779876 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.784857 4706 policy_none.go:49] "None policy: Start" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.785826 4706 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.785852 4706 state_mem.go:35] "Initializing new in-memory state store" Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.801016 4706 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.857522 4706 manager.go:334] "Starting Device Plugin manager" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.857600 4706 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.857614 4706 server.go:79] "Starting device plugin registration server" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.858067 4706 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.858084 4706 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.858263 4706 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.858467 4706 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.858491 4706 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.871972 4706 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.876533 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.876625 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.877797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.877832 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.877844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.878263 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.878305 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.878643 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.879138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.879187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.879204 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.880648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.880685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.880703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.880883 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.881152 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.881260 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.881960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.882029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.882049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.882382 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.882539 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.882998 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.883303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.883344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.883354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.884774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.884797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.884817 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.885067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.885144 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.885184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.885545 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.886553 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.886620 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.887497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.887528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.887549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.888009 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.888142 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.888185 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.888327 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.888416 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.891038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.891061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.891074 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.907898 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="400ms" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937074 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937122 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937153 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937178 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937203 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937326 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937387 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937428 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937514 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937541 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937564 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937593 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937614 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937638 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.937704 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.958469 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.959672 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.959790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.959878 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:36 crc kubenswrapper[4706]: I1127 07:08:36.959968 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 07:08:36 crc kubenswrapper[4706]: E1127 07:08:36.960602 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038786 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038839 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038856 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038874 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038892 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038908 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038923 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038945 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038962 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038976 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.038990 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039005 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039021 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039036 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039052 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039403 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039461 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039466 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039518 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039449 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039530 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039501 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039552 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039553 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039600 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039572 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039586 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039588 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039604 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.039434 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.161615 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.163206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.163354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.163389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.163448 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 07:08:37 crc kubenswrapper[4706]: E1127 07:08:37.164102 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.222056 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.234161 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.251750 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.261404 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.266310 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:37 crc kubenswrapper[4706]: W1127 07:08:37.268765 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-2ff98e0b8bae0c85f6b569e1be83f4106880f908f23eb5c6b9894627c4ffa219 WatchSource:0}: Error finding container 2ff98e0b8bae0c85f6b569e1be83f4106880f908f23eb5c6b9894627c4ffa219: Status 404 returned error can't find the container with id 2ff98e0b8bae0c85f6b569e1be83f4106880f908f23eb5c6b9894627c4ffa219 Nov 27 07:08:37 crc kubenswrapper[4706]: W1127 07:08:37.281116 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-aea026e632b97d90e5aabdefc85c48738a77c72297defbb5496f1f3153378e85 WatchSource:0}: Error finding container aea026e632b97d90e5aabdefc85c48738a77c72297defbb5496f1f3153378e85: Status 404 returned error can't find the container with id aea026e632b97d90e5aabdefc85c48738a77c72297defbb5496f1f3153378e85 Nov 27 07:08:37 crc kubenswrapper[4706]: W1127 07:08:37.287792 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a34b7813ba80e735829af41cd11cd298cb1e3fc2a41129ded6eb4c8e2e84528a WatchSource:0}: Error finding container a34b7813ba80e735829af41cd11cd298cb1e3fc2a41129ded6eb4c8e2e84528a: Status 404 returned error can't find the container with id a34b7813ba80e735829af41cd11cd298cb1e3fc2a41129ded6eb4c8e2e84528a Nov 27 07:08:37 crc kubenswrapper[4706]: W1127 07:08:37.288978 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e9a69d9b3fd24a57236506c679f144c7378082a13944217c420fef05a83493d0 WatchSource:0}: Error finding container e9a69d9b3fd24a57236506c679f144c7378082a13944217c420fef05a83493d0: Status 404 returned error can't find the container with id e9a69d9b3fd24a57236506c679f144c7378082a13944217c420fef05a83493d0 Nov 27 07:08:37 crc kubenswrapper[4706]: E1127 07:08:37.309434 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="800ms" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.565106 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.568586 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.568655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.568672 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.568717 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 07:08:37 crc kubenswrapper[4706]: E1127 07:08:37.569755 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Nov 27 07:08:37 crc kubenswrapper[4706]: W1127 07:08:37.611191 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:37 crc kubenswrapper[4706]: E1127 07:08:37.611368 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:37 crc kubenswrapper[4706]: W1127 07:08:37.638848 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:37 crc kubenswrapper[4706]: E1127 07:08:37.638985 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:37 crc kubenswrapper[4706]: E1127 07:08:37.651939 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.153:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187bcb65b2b91a28 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 07:08:36.69629188 +0000 UTC m=+0.585882730,LastTimestamp:2025-11-27 07:08:36.69629188 +0000 UTC m=+0.585882730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.698033 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.701105 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 23:26:04.759358187 +0000 UTC Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.782373 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a34b7813ba80e735829af41cd11cd298cb1e3fc2a41129ded6eb4c8e2e84528a"} Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.783847 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"aea026e632b97d90e5aabdefc85c48738a77c72297defbb5496f1f3153378e85"} Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.786792 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2ff98e0b8bae0c85f6b569e1be83f4106880f908f23eb5c6b9894627c4ffa219"} Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.788916 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e2cb4fcbb27d7aba2ec081760884070006d35c60220d32a7d7a21066ac71e2a0"} Nov 27 07:08:37 crc kubenswrapper[4706]: W1127 07:08:37.790275 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:37 crc kubenswrapper[4706]: E1127 07:08:37.790437 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:37 crc kubenswrapper[4706]: I1127 07:08:37.790795 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e9a69d9b3fd24a57236506c679f144c7378082a13944217c420fef05a83493d0"} Nov 27 07:08:38 crc kubenswrapper[4706]: E1127 07:08:38.110933 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="1.6s" Nov 27 07:08:38 crc kubenswrapper[4706]: W1127 07:08:38.288855 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:38 crc kubenswrapper[4706]: E1127 07:08:38.288979 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.370388 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.371976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.372055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.372079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.372136 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 07:08:38 crc kubenswrapper[4706]: E1127 07:08:38.373054 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.698163 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.701216 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 02:49:48.837729079 +0000 UTC Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.701299 4706 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1147h41m10.136432903s for next certificate rotation Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.798091 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ba7abfb64fc109422ff1701b203d9c4b150af7c5c4ab3693a13170d9c88e3e8d" exitCode=0 Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.798206 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ba7abfb64fc109422ff1701b203d9c4b150af7c5c4ab3693a13170d9c88e3e8d"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.798318 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.800018 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.800080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.800103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.802519 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f6c992ab3d6e0b48aacf3fbe205d19319b19811aecef632383af133699738bb8" exitCode=0 Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.802696 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f6c992ab3d6e0b48aacf3fbe205d19319b19811aecef632383af133699738bb8"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.802762 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.805453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.805497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.805515 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.809647 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.809731 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.809768 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.809661 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.809788 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.811270 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.811335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.811358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.812476 4706 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c" exitCode=0 Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.812609 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.812611 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.813768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.813803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.813823 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.817117 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159" exitCode=0 Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.817194 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159"} Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.817285 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.818323 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.818376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.818398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.818960 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.819496 4706 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": dial tcp 192.168.126.11:10357: connect: connection refused" start-of-body= Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.819566 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": dial tcp 192.168.126.11:10357: connect: connection refused" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.820737 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.826562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.826621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:38 crc kubenswrapper[4706]: I1127 07:08:38.826645 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.698119 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:39 crc kubenswrapper[4706]: E1127 07:08:39.712039 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="3.2s" Nov 27 07:08:39 crc kubenswrapper[4706]: W1127 07:08:39.820664 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.153:6443: connect: connection refused Nov 27 07:08:39 crc kubenswrapper[4706]: E1127 07:08:39.820785 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.153:6443: connect: connection refused" logger="UnhandledError" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.823943 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"13c69b1e609830c4997b55f75cfc8912829a78f8909f6cd895915af0f71c0859"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.824096 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.825508 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.825554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.825572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.829044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.829113 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.829133 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.829234 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.830703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.830766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.830778 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.833809 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.833876 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.833901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.833920 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.835334 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ed4e993b57709ab51cce04bb7fca3248379da5781fe683d97b95e8426519904f" exitCode=0 Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.835455 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.835465 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.836108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ed4e993b57709ab51cce04bb7fca3248379da5781fe683d97b95e8426519904f"} Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.837427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.837464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.837477 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.838342 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.838378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.838390 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.973442 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.975168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.975216 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.975228 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:39 crc kubenswrapper[4706]: I1127 07:08:39.975288 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 07:08:39 crc kubenswrapper[4706]: E1127 07:08:39.975965 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.153:6443: connect: connection refused" node="crc" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.843615 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116"} Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.843674 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.845626 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.845687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.845707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.846621 4706 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="db84314fb0a0ca52f942e902f1ff40f3c35c13c9f13b9731095cfc8616365270" exitCode=0 Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.846774 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.846784 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.846826 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.846790 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"db84314fb0a0ca52f942e902f1ff40f3c35c13c9f13b9731095cfc8616365270"} Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.846778 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.846974 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848353 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:40 crc kubenswrapper[4706]: I1127 07:08:40.848603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.853902 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"234b8b6614e3e6bc4ebeee44fdf873b47b1d0243e2de0c544d0b66dbf81a09f5"} Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.853985 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e2b8eadd7c4316cae5228c9b5cc34aae24214678670204bbbc7a6702ba897de7"} Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.854006 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"816d3417d6551afcad4145fe4eced85a163aef98d3a55397ca2b051b0d78bd1b"} Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.854017 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.854024 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"55fd41597e49d6926dd70a076f356cf3bd95e8af990977f71d9cd6ccc52fb4f8"} Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.854171 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.854945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.854994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:41 crc kubenswrapper[4706]: I1127 07:08:41.855007 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.864291 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"adbc5d4385eb4b49090b2484b8328ecb677c2ca4c1209b4d34d6480f9cf6b07c"} Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.864382 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.864516 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.865551 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.865591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.865603 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.865780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.865846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:42 crc kubenswrapper[4706]: I1127 07:08:42.865869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.039274 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.176528 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.178475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.178539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.178557 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.178597 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.181701 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.319950 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.320206 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.321750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.321789 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.321801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.867856 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.868040 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.869151 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.869214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.869274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.869730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.869774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:43 crc kubenswrapper[4706]: I1127 07:08:43.869788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.206431 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.342493 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.342744 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.344570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.344631 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.344646 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.763165 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.870700 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.870878 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.871734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.871779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.871790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.872360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.872386 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:44 crc kubenswrapper[4706]: I1127 07:08:44.872396 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:45 crc kubenswrapper[4706]: I1127 07:08:45.874277 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:45 crc kubenswrapper[4706]: I1127 07:08:45.875563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:45 crc kubenswrapper[4706]: I1127 07:08:45.875606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:45 crc kubenswrapper[4706]: I1127 07:08:45.875618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:46 crc kubenswrapper[4706]: E1127 07:08:46.872900 4706 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.107371 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.107566 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.108624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.108665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.108674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.113046 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.878543 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.878800 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.880655 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.880710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.880728 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:47 crc kubenswrapper[4706]: I1127 07:08:47.883482 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:48 crc kubenswrapper[4706]: I1127 07:08:48.880766 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:48 crc kubenswrapper[4706]: I1127 07:08:48.881800 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:48 crc kubenswrapper[4706]: I1127 07:08:48.881854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:48 crc kubenswrapper[4706]: I1127 07:08:48.881869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:49 crc kubenswrapper[4706]: I1127 07:08:49.883209 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:49 crc kubenswrapper[4706]: I1127 07:08:49.884276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:49 crc kubenswrapper[4706]: I1127 07:08:49.884321 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:49 crc kubenswrapper[4706]: I1127 07:08:49.884341 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:50 crc kubenswrapper[4706]: W1127 07:08:50.326214 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 27 07:08:50 crc kubenswrapper[4706]: I1127 07:08:50.326356 4706 trace.go:236] Trace[46079394]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 07:08:40.324) (total time: 10001ms): Nov 27 07:08:50 crc kubenswrapper[4706]: Trace[46079394]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:08:50.326) Nov 27 07:08:50 crc kubenswrapper[4706]: Trace[46079394]: [10.001669236s] [10.001669236s] END Nov 27 07:08:50 crc kubenswrapper[4706]: E1127 07:08:50.326386 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 27 07:08:50 crc kubenswrapper[4706]: W1127 07:08:50.376410 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 27 07:08:50 crc kubenswrapper[4706]: I1127 07:08:50.376509 4706 trace.go:236] Trace[151391630]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 07:08:40.375) (total time: 10000ms): Nov 27 07:08:50 crc kubenswrapper[4706]: Trace[151391630]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (07:08:50.376) Nov 27 07:08:50 crc kubenswrapper[4706]: Trace[151391630]: [10.000936151s] [10.000936151s] END Nov 27 07:08:50 crc kubenswrapper[4706]: E1127 07:08:50.376529 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 27 07:08:50 crc kubenswrapper[4706]: W1127 07:08:50.633068 4706 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 27 07:08:50 crc kubenswrapper[4706]: I1127 07:08:50.633159 4706 trace.go:236] Trace[1329179642]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 07:08:40.631) (total time: 10001ms): Nov 27 07:08:50 crc kubenswrapper[4706]: Trace[1329179642]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:08:50.633) Nov 27 07:08:50 crc kubenswrapper[4706]: Trace[1329179642]: [10.001498439s] [10.001498439s] END Nov 27 07:08:50 crc kubenswrapper[4706]: E1127 07:08:50.633186 4706 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 27 07:08:50 crc kubenswrapper[4706]: I1127 07:08:50.699547 4706 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 27 07:08:51 crc kubenswrapper[4706]: I1127 07:08:51.732729 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 27 07:08:51 crc kubenswrapper[4706]: I1127 07:08:51.732800 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 07:08:51 crc kubenswrapper[4706]: I1127 07:08:51.739751 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 27 07:08:51 crc kubenswrapper[4706]: I1127 07:08:51.739827 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 07:08:51 crc kubenswrapper[4706]: I1127 07:08:51.819982 4706 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 07:08:51 crc kubenswrapper[4706]: I1127 07:08:51.820095 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.061420 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.061612 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.062637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.062671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.062682 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.074448 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.890863 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.891576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.891607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:53 crc kubenswrapper[4706]: I1127 07:08:53.891616 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.211545 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.211784 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.213035 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.213092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.213111 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.217797 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.893925 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.894008 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.895192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.895244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:08:54 crc kubenswrapper[4706]: I1127 07:08:54.895254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.128540 4706 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.372416 4706 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.421746 4706 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.693323 4706 apiserver.go:52] "Watching apiserver" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.698298 4706 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.698561 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.698887 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.699789 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.700041 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.700702 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.700705 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.700779 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.700851 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.702179 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.702407 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.702939 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.703190 4706 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.704437 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.704858 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.706664 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.706688 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.706663 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.706731 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.706729 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.707152 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.716988 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.721531 4706 trace.go:236] Trace[1906158358]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 07:08:43.743) (total time: 12978ms): Nov 27 07:08:56 crc kubenswrapper[4706]: Trace[1906158358]: ---"Objects listed" error: 12978ms (07:08:56.721) Nov 27 07:08:56 crc kubenswrapper[4706]: Trace[1906158358]: [12.978383657s] [12.978383657s] END Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.721567 4706 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.721744 4706 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.722336 4706 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.727951 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.738842 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.754169 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.760462 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47536->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.760549 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47536->192.168.126.11:17697: read: connection reset by peer" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.760511 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47534->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.760599 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47534->192.168.126.11:17697: read: connection reset by peer" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.760942 4706 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.761012 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.763184 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.772812 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.786437 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.797238 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.805379 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.819199 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822508 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822553 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822614 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822631 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822647 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822663 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822678 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822693 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822708 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822724 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822741 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822757 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822772 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822788 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822804 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822822 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822822 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822831 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822842 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822878 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822897 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822917 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822932 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822951 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822965 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822979 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822981 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.822994 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823009 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823025 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823043 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823057 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823071 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823088 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823105 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823113 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823126 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823138 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823141 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823173 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823188 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823202 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823247 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823280 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823301 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823321 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823345 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823369 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823391 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823413 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823436 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823458 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823481 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823504 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823527 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823549 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823576 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823602 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823627 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823658 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823681 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823704 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823725 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823746 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823771 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823794 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823916 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823945 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824037 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824065 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824087 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824110 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824133 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824161 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824184 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824207 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824250 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824276 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824335 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824363 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824386 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824410 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824434 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824456 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824478 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824499 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824521 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824541 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824564 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824587 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824614 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824639 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824663 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824688 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824713 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824737 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824762 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824787 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824814 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824838 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824860 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824888 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824914 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824938 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824961 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824985 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825008 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825030 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825050 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825073 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823320 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823698 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823717 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823899 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823984 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.823995 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824533 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824625 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824906 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.824932 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825057 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825082 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825090 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825150 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825108 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825329 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825343 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825350 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825374 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825405 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825415 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825436 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825461 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825486 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825508 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825534 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825542 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825559 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825589 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825617 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825644 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825668 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825691 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825713 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826079 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826109 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826127 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826148 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826181 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826211 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826251 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826270 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826291 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826311 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826331 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826348 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826366 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826513 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826536 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826553 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827052 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827074 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827357 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827384 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827441 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827459 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828687 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828720 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828766 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828788 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828815 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828863 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828883 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828901 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828940 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828960 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828978 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829018 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829037 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829164 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829189 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829213 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829257 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829274 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829293 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829332 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829352 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829370 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829408 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829428 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829446 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829481 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829502 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829518 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829553 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829574 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829591 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829607 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829642 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830801 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830825 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830844 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830882 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830899 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830917 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830954 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830972 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830990 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831027 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831049 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831069 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831104 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831123 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831140 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831159 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831194 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831212 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831279 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831321 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831399 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831429 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831448 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831491 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831512 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831536 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831578 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831598 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831633 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831706 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831807 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825508 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.825623 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826504 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826524 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826674 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826899 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.826998 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827110 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827120 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827188 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827324 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827524 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827594 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827800 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827818 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827871 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827902 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.827999 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828008 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828106 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828137 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828206 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828441 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828449 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828456 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828488 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828510 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828539 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828674 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.828725 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829019 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829144 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829492 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829636 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832679 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829690 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829829 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829924 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.829943 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830362 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830638 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830720 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830784 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830903 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.830986 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831126 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831131 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831218 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831378 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831451 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831457 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831480 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831749 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.831998 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832842 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832936 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832245 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832255 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832259 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833589 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832455 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832833 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833168 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833327 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833380 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833456 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833510 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833607 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833706 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833720 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.833826 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.834007 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:08:57.333987682 +0000 UTC m=+21.223578492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.835862 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.835975 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.836321 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.836470 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.836505 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.836635 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.836722 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.836816 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.837096 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.837207 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.837409 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.837746 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.837780 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.838065 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.838117 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.838447 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.838672 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.838938 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839005 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839053 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.832453 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839259 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839347 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839360 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839370 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839380 4706 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839391 4706 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839403 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839413 4706 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839424 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839434 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839444 4706 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839454 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839465 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839474 4706 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839467 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839484 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839528 4706 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839545 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839561 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839578 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839640 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839657 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839577 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839672 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839721 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839736 4706 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839748 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839764 4706 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839782 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839797 4706 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839811 4706 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839824 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839838 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839849 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839861 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839871 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839880 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839891 4706 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839900 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839912 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839922 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839921 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839933 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839969 4706 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.839985 4706 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840000 4706 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840013 4706 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840026 4706 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840049 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841785 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841813 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841830 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841845 4706 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841858 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841872 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841886 4706 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841897 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841938 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841950 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841965 4706 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841979 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841992 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842005 4706 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842017 4706 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842037 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842054 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842069 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842079 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842088 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842089 4706 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842097 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842362 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842381 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842396 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842408 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842420 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842432 4706 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842444 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842455 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842467 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842479 4706 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842490 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842502 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842513 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842564 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840248 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840652 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840668 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840851 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841146 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841163 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.840773 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841372 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841499 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841648 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841614 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842626 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841735 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841949 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842900 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842914 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842926 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843080 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841973 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.841981 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842031 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842264 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842220 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842358 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843220 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.842395 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843303 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.843324 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843348 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.843483 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:57.343426843 +0000 UTC m=+21.233017763 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843509 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.843547 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843570 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843586 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.843626 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:57.343616269 +0000 UTC m=+21.233207169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843640 4706 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843652 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843741 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843756 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843768 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843780 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843793 4706 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843805 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843818 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843831 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843844 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843856 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843872 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843884 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843898 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843910 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843924 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843937 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843949 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843962 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843975 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843988 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844000 4706 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844013 4706 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844024 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844040 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844052 4706 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844064 4706 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.843903 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844008 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844135 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844282 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844407 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844543 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844714 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844717 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.844835 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.845129 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.845254 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.845833 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.846375 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.846393 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.846749 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.847198 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.847534 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.847610 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.847786 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.848277 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.855943 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.856180 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.856210 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.856176 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.856254 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.856267 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.856483 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:57.356303619 +0000 UTC m=+21.245894549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.856672 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.857252 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.857313 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.857338 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.857408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.857447 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.858883 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.859452 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.859630 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.859651 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.859660 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.859630 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: E1127 07:08:56.859745 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:57.359733435 +0000 UTC m=+21.249324235 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.859828 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.859966 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.859889 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.861702 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.863651 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.864439 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.868025 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.869127 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.869576 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.889755 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.891418 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.891793 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.891969 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.892025 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.892108 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.892359 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.892375 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.892420 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.892761 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.893195 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.893251 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.893556 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.893763 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.897972 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.900589 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.900613 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.901989 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116" exitCode=255 Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.902021 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116"} Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.904079 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.908565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.912536 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.915009 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.915088 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.915188 4706 scope.go:117] "RemoveContainer" containerID="8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.917457 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.924798 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.933718 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.943250 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.944857 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.944932 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.944998 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.945343 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.945336 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.945909 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.945942 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.945957 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.945977 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.945994 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946006 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946018 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946030 4706 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946041 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946057 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946068 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946081 4706 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946096 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946112 4706 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946131 4706 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946143 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946156 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946169 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946182 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946197 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946209 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946246 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946261 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946273 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946286 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946298 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946311 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946327 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946346 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946361 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946380 4706 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946396 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946409 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946423 4706 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946435 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946449 4706 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946463 4706 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946482 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946497 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946535 4706 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946555 4706 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946573 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946587 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946600 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946613 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946627 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946640 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946652 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946666 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946679 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946693 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946707 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946720 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946735 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946747 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946758 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946770 4706 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946783 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946795 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946807 4706 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946819 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946832 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946844 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946857 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946869 4706 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946882 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946900 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946913 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946926 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946939 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946951 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.946964 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947011 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947025 4706 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947038 4706 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947051 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947064 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947077 4706 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947090 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947106 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.947126 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.956235 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:56 crc kubenswrapper[4706]: I1127 07:08:56.966455 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.015165 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.021188 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 07:08:57 crc kubenswrapper[4706]: W1127 07:08:57.027365 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-cba18a0bfb13974abbdfa205781eba9dba7a0eff29748805f8799f2d047f0358 WatchSource:0}: Error finding container cba18a0bfb13974abbdfa205781eba9dba7a0eff29748805f8799f2d047f0358: Status 404 returned error can't find the container with id cba18a0bfb13974abbdfa205781eba9dba7a0eff29748805f8799f2d047f0358 Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.029180 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 07:08:57 crc kubenswrapper[4706]: W1127 07:08:57.048127 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-cd12802aebfa8c4c036a0366252daa06dae0770995129a4e0d973f41db2e8fa3 WatchSource:0}: Error finding container cd12802aebfa8c4c036a0366252daa06dae0770995129a4e0d973f41db2e8fa3: Status 404 returned error can't find the container with id cd12802aebfa8c4c036a0366252daa06dae0770995129a4e0d973f41db2e8fa3 Nov 27 07:08:57 crc kubenswrapper[4706]: W1127 07:08:57.048482 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-9f62c04abde6257be253c7708263f112333b5fa302007a3f634d520850958531 WatchSource:0}: Error finding container 9f62c04abde6257be253c7708263f112333b5fa302007a3f634d520850958531: Status 404 returned error can't find the container with id 9f62c04abde6257be253c7708263f112333b5fa302007a3f634d520850958531 Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.349960 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.350045 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.350065 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.350153 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:08:58.350124331 +0000 UTC m=+22.239715141 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.350166 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.350195 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.350240 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:58.350208974 +0000 UTC m=+22.239799784 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.350312 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:58.350287446 +0000 UTC m=+22.239878316 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.450390 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.450447 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450548 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450563 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450574 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450624 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:58.450608582 +0000 UTC m=+22.340199392 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450689 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450726 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450740 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:57 crc kubenswrapper[4706]: E1127 07:08:57.450823 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:08:58.450793348 +0000 UTC m=+22.340384228 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.905729 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"cd12802aebfa8c4c036a0366252daa06dae0770995129a4e0d973f41db2e8fa3"} Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.907846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee"} Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.907906 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e"} Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.907920 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9f62c04abde6257be253c7708263f112333b5fa302007a3f634d520850958531"} Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.909965 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010"} Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.910008 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cba18a0bfb13974abbdfa205781eba9dba7a0eff29748805f8799f2d047f0358"} Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.912024 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.913826 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c"} Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.914178 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.925862 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.937371 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.948136 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.963866 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.978438 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:57 crc kubenswrapper[4706]: I1127 07:08:57.990541 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.003457 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.017920 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.033904 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.058558 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.085864 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.100618 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.116642 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.130912 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.240376 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5mrfn"] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.240763 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-c44hv"] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.240960 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.241005 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-m645h"] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.241203 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.241792 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.244929 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245009 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245140 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245359 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245439 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245527 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245586 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245601 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245659 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245756 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.245796 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.246666 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.250160 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.258791 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.271407 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.286931 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.299988 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.319116 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.335771 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.356744 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.356858 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.356885 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.356914 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-k8s-cni-cncf-io\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.356943 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:09:00.356912073 +0000 UTC m=+24.246502883 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.356990 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-kubelet\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357020 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-etc-kubernetes\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.357041 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.357042 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4272330e-1068-42ec-92d1-3a3503adaef1-hosts-file\") pod \"node-resolver-5mrfn\" (UID: \"4272330e-1068-42ec-92d1-3a3503adaef1\") " pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.357135 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:00.357114429 +0000 UTC m=+24.246705309 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.357164 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:00.35715388 +0000 UTC m=+24.246744690 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357167 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8hdn\" (UniqueName: \"kubernetes.io/projected/4272330e-1068-42ec-92d1-3a3503adaef1-kube-api-access-c8hdn\") pod \"node-resolver-5mrfn\" (UID: \"4272330e-1068-42ec-92d1-3a3503adaef1\") " pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357201 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-multus-certs\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357253 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f89d12ee-9aae-48da-a7a6-130c7186f506-proxy-tls\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357273 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-cni-bin\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357290 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-hostroot\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357350 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0e33550a-5308-4e22-9c81-8a12e97f44ef-cni-binary-copy\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357370 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-netns\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357386 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv675\" (UniqueName: \"kubernetes.io/projected/f89d12ee-9aae-48da-a7a6-130c7186f506-kube-api-access-sv675\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357408 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-cni-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357427 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-cni-multus\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357781 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-daemon-config\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f89d12ee-9aae-48da-a7a6-130c7186f506-mcd-auth-proxy-config\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357840 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f89d12ee-9aae-48da-a7a6-130c7186f506-rootfs\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357869 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-conf-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357898 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-cnibin\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357923 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-socket-dir-parent\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357943 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ns2v\" (UniqueName: \"kubernetes.io/projected/0e33550a-5308-4e22-9c81-8a12e97f44ef-kube-api-access-6ns2v\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.357962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-system-cni-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.358047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-os-release\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.378438 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.405288 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.437971 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459309 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-cni-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459370 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-cni-multus\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459393 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-daemon-config\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459416 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f89d12ee-9aae-48da-a7a6-130c7186f506-mcd-auth-proxy-config\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459438 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-conf-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459459 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f89d12ee-9aae-48da-a7a6-130c7186f506-rootfs\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459495 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-cnibin\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459518 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ns2v\" (UniqueName: \"kubernetes.io/projected/0e33550a-5308-4e22-9c81-8a12e97f44ef-kube-api-access-6ns2v\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459533 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-cni-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459590 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-conf-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459594 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-socket-dir-parent\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459539 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-socket-dir-parent\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459626 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-os-release\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459636 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f89d12ee-9aae-48da-a7a6-130c7186f506-rootfs\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459647 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-system-cni-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459676 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-cnibin\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459688 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-k8s-cni-cncf-io\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459708 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-kubelet\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459727 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-etc-kubernetes\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459748 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4272330e-1068-42ec-92d1-3a3503adaef1-hosts-file\") pod \"node-resolver-5mrfn\" (UID: \"4272330e-1068-42ec-92d1-3a3503adaef1\") " pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459769 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8hdn\" (UniqueName: \"kubernetes.io/projected/4272330e-1068-42ec-92d1-3a3503adaef1-kube-api-access-c8hdn\") pod \"node-resolver-5mrfn\" (UID: \"4272330e-1068-42ec-92d1-3a3503adaef1\") " pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459787 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-hostroot\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459806 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-multus-certs\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459824 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f89d12ee-9aae-48da-a7a6-130c7186f506-proxy-tls\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459845 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-cni-bin\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459896 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459914 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0e33550a-5308-4e22-9c81-8a12e97f44ef-cni-binary-copy\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459930 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-netns\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459947 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv675\" (UniqueName: \"kubernetes.io/projected/f89d12ee-9aae-48da-a7a6-130c7186f506-kube-api-access-sv675\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.459973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-hostroot\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460156 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0e33550a-5308-4e22-9c81-8a12e97f44ef-multus-daemon-config\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460177 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-multus-certs\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460164 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-os-release\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460197 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-etc-kubernetes\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460253 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-cni-bin\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460260 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-cni-multus\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460269 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-system-cni-dir\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460330 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460346 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4272330e-1068-42ec-92d1-3a3503adaef1-hosts-file\") pod \"node-resolver-5mrfn\" (UID: \"4272330e-1068-42ec-92d1-3a3503adaef1\") " pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460349 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460356 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f89d12ee-9aae-48da-a7a6-130c7186f506-mcd-auth-proxy-config\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460371 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-netns\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460375 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460448 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-run-k8s-cni-cncf-io\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460484 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e33550a-5308-4e22-9c81-8a12e97f44ef-host-var-lib-kubelet\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460499 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:00.460477659 +0000 UTC m=+24.350068569 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460423 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460548 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460562 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.460593 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:00.460585002 +0000 UTC m=+24.350175932 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.460752 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0e33550a-5308-4e22-9c81-8a12e97f44ef-cni-binary-copy\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.466945 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f89d12ee-9aae-48da-a7a6-130c7186f506-proxy-tls\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.482841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ns2v\" (UniqueName: \"kubernetes.io/projected/0e33550a-5308-4e22-9c81-8a12e97f44ef-kube-api-access-6ns2v\") pod \"multus-m645h\" (UID: \"0e33550a-5308-4e22-9c81-8a12e97f44ef\") " pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.486951 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv675\" (UniqueName: \"kubernetes.io/projected/f89d12ee-9aae-48da-a7a6-130c7186f506-kube-api-access-sv675\") pod \"machine-config-daemon-c44hv\" (UID: \"f89d12ee-9aae-48da-a7a6-130c7186f506\") " pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.492112 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.493345 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8hdn\" (UniqueName: \"kubernetes.io/projected/4272330e-1068-42ec-92d1-3a3503adaef1-kube-api-access-c8hdn\") pod \"node-resolver-5mrfn\" (UID: \"4272330e-1068-42ec-92d1-3a3503adaef1\") " pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.505140 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.518858 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.529483 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.541164 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.550388 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.553290 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5mrfn" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.561948 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:08:58 crc kubenswrapper[4706]: W1127 07:08:58.565272 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4272330e_1068_42ec_92d1_3a3503adaef1.slice/crio-cafe48cd9bf415176aa3a2f9a54956ac1046adb81f094990f147f22621908591 WatchSource:0}: Error finding container cafe48cd9bf415176aa3a2f9a54956ac1046adb81f094990f147f22621908591: Status 404 returned error can't find the container with id cafe48cd9bf415176aa3a2f9a54956ac1046adb81f094990f147f22621908591 Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.565509 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.566877 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m645h" Nov 27 07:08:58 crc kubenswrapper[4706]: W1127 07:08:58.573692 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf89d12ee_9aae_48da_a7a6_130c7186f506.slice/crio-46bc04a7669d744d44f1fbd28cc4a14116e6ec47f42e170d5369b043e83d0621 WatchSource:0}: Error finding container 46bc04a7669d744d44f1fbd28cc4a14116e6ec47f42e170d5369b043e83d0621: Status 404 returned error can't find the container with id 46bc04a7669d744d44f1fbd28cc4a14116e6ec47f42e170d5369b043e83d0621 Nov 27 07:08:58 crc kubenswrapper[4706]: W1127 07:08:58.580445 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e33550a_5308_4e22_9c81_8a12e97f44ef.slice/crio-1c9598da954835dc9e28dfeffc0536b04e23b4275a9bfcf2a7af5b5110f942d7 WatchSource:0}: Error finding container 1c9598da954835dc9e28dfeffc0536b04e23b4275a9bfcf2a7af5b5110f942d7: Status 404 returned error can't find the container with id 1c9598da954835dc9e28dfeffc0536b04e23b4275a9bfcf2a7af5b5110f942d7 Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.580768 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.598937 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.632283 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-p9dt7"] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.633762 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhhb"] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.634025 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.635378 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.637080 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638146 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638182 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638359 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638417 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638484 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638678 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638870 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.638927 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.654640 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.669725 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.682075 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.693502 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.707538 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.728778 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.747522 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.761089 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762458 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-slash\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762492 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-systemd\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762508 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-netns\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762541 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s94lx\" (UniqueName: \"kubernetes.io/projected/d1b93703-3a71-49b6-bff6-b4d314006ddd-kube-api-access-s94lx\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762560 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-os-release\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762576 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-script-lib\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762619 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8cb8d10b-fddc-42e4-9181-6366603ea340-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762634 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-ovn-kubernetes\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762648 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-bin\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762689 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-kubelet\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762703 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-etc-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762717 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-node-log\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762731 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-config\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762766 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762783 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-ovn\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762797 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-systemd-units\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-system-cni-dir\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762912 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-var-lib-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762930 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-log-socket\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762947 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762964 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-netd\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.762997 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.763021 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovn-node-metrics-cert\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.763037 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlfzt\" (UniqueName: \"kubernetes.io/projected/8cb8d10b-fddc-42e4-9181-6366603ea340-kube-api-access-jlfzt\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.763070 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-env-overrides\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.763087 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-cnibin\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.763105 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8cb8d10b-fddc-42e4-9181-6366603ea340-cni-binary-copy\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.775813 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.775839 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.775958 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.775830 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.776087 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.776137 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.777173 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.780190 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.780750 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.781587 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.782176 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.782809 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.784362 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.784919 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.785496 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.786539 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.787053 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.787957 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.788688 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.789621 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.790109 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.791074 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.791316 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.791614 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.792176 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.793036 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.793593 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.794635 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.795082 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.795638 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.796473 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.797074 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.797878 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.798486 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.799535 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.800173 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.801389 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.802090 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.802825 4706 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.802935 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.806809 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.807414 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.808327 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.810131 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.810859 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.811823 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.812480 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.813544 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.814033 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.814695 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.815755 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.816725 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.817214 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.818124 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.818687 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.819863 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.819853 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.820374 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.821189 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.821708 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.822248 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.823147 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.823704 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.824629 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.827744 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.834794 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.835545 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.848071 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.861477 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863775 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-ovn\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863803 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-systemd-units\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863828 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-system-cni-dir\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863877 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-var-lib-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863879 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-ovn\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863894 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-log-socket\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863943 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-log-socket\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863966 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863989 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-netd\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864009 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-system-cni-dir\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864043 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864069 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-netd\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.863989 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-systemd-units\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864089 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864010 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864122 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovn-node-metrics-cert\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864138 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlfzt\" (UniqueName: \"kubernetes.io/projected/8cb8d10b-fddc-42e4-9181-6366603ea340-kube-api-access-jlfzt\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864157 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-env-overrides\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864172 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-cnibin\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864187 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8cb8d10b-fddc-42e4-9181-6366603ea340-cni-binary-copy\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864204 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-slash\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864236 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-systemd\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864253 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-netns\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864271 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s94lx\" (UniqueName: \"kubernetes.io/projected/d1b93703-3a71-49b6-bff6-b4d314006ddd-kube-api-access-s94lx\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864286 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-os-release\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864301 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-script-lib\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864323 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8cb8d10b-fddc-42e4-9181-6366603ea340-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864339 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-ovn-kubernetes\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864354 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-bin\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864374 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-kubelet\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864390 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-etc-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-node-log\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864418 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-config\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864439 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864502 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864104 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-var-lib-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.864789 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-os-release\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-kubelet\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865044 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-node-log\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865110 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-ovn-kubernetes\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865140 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-bin\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865168 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-etc-openvswitch\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865201 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-slash\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865246 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8cb8d10b-fddc-42e4-9181-6366603ea340-cnibin\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865320 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-netns\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865327 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-systemd\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865342 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-env-overrides\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865399 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-script-lib\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.865758 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8cb8d10b-fddc-42e4-9181-6366603ea340-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.866055 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-config\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.866176 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8cb8d10b-fddc-42e4-9181-6366603ea340-cni-binary-copy\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.869311 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovn-node-metrics-cert\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.878198 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlfzt\" (UniqueName: \"kubernetes.io/projected/8cb8d10b-fddc-42e4-9181-6366603ea340-kube-api-access-jlfzt\") pod \"multus-additional-cni-plugins-p9dt7\" (UID: \"8cb8d10b-fddc-42e4-9181-6366603ea340\") " pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.879547 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s94lx\" (UniqueName: \"kubernetes.io/projected/d1b93703-3a71-49b6-bff6-b4d314006ddd-kube-api-access-s94lx\") pod \"ovnkube-node-7rhhb\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.879782 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.890959 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.901134 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.911470 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.917059 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5mrfn" event={"ID":"4272330e-1068-42ec-92d1-3a3503adaef1","Type":"ContainerStarted","Data":"790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.917297 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5mrfn" event={"ID":"4272330e-1068-42ec-92d1-3a3503adaef1","Type":"ContainerStarted","Data":"cafe48cd9bf415176aa3a2f9a54956ac1046adb81f094990f147f22621908591"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.918921 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.920739 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.920782 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.920796 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"46bc04a7669d744d44f1fbd28cc4a14116e6ec47f42e170d5369b043e83d0621"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.922486 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerStarted","Data":"f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.922507 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerStarted","Data":"1c9598da954835dc9e28dfeffc0536b04e23b4275a9bfcf2a7af5b5110f942d7"} Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.927907 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: E1127 07:08:58.928104 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.941944 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.948694 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.955699 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.961984 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: W1127 07:08:58.966126 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cb8d10b_fddc_42e4_9181_6366603ea340.slice/crio-44c749c5fa0e7a82ca2e75e74844d0b3491303bd47ba2d2f1e4e01d017b0d210 WatchSource:0}: Error finding container 44c749c5fa0e7a82ca2e75e74844d0b3491303bd47ba2d2f1e4e01d017b0d210: Status 404 returned error can't find the container with id 44c749c5fa0e7a82ca2e75e74844d0b3491303bd47ba2d2f1e4e01d017b0d210 Nov 27 07:08:58 crc kubenswrapper[4706]: W1127 07:08:58.970887 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b93703_3a71_49b6_bff6_b4d314006ddd.slice/crio-855a8b79edc691d8ad0e3d12d589287fd4cca9ede85f6439ef7b744883fabcd9 WatchSource:0}: Error finding container 855a8b79edc691d8ad0e3d12d589287fd4cca9ede85f6439ef7b744883fabcd9: Status 404 returned error can't find the container with id 855a8b79edc691d8ad0e3d12d589287fd4cca9ede85f6439ef7b744883fabcd9 Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.975715 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:58 crc kubenswrapper[4706]: I1127 07:08:58.990867 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:58Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.047024 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.072590 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.115737 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.134473 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.167728 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.182550 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.197307 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.216194 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.228023 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.240728 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.251837 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.271168 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.285899 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.926593 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172" exitCode=0 Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.926670 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.927211 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"855a8b79edc691d8ad0e3d12d589287fd4cca9ede85f6439ef7b744883fabcd9"} Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.929797 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cb8d10b-fddc-42e4-9181-6366603ea340" containerID="98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2" exitCode=0 Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.929856 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerDied","Data":"98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2"} Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.931091 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerStarted","Data":"44c749c5fa0e7a82ca2e75e74844d0b3491303bd47ba2d2f1e4e01d017b0d210"} Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.943917 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.957889 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.969902 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.982461 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:08:59 crc kubenswrapper[4706]: I1127 07:08:59.998266 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:08:59Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.011054 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.023538 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.035729 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.045768 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.060005 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.073469 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.088201 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.105627 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.121903 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.135423 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.150888 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.164198 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.182443 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.195937 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.211286 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.224375 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.236035 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.248572 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.261242 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.274457 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.295067 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.380290 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.380454 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.380497 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:09:04.380460764 +0000 UTC m=+28.270051564 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.380549 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.380588 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.380640 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:04.380625609 +0000 UTC m=+28.270216409 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.380656 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.380695 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:04.380686001 +0000 UTC m=+28.270276811 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.481261 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.481306 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481452 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481486 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481500 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481456 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481589 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481602 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481556 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:04.481533863 +0000 UTC m=+28.371124733 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.481653 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:04.481641266 +0000 UTC m=+28.371232366 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.776366 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.776389 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.776792 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.776508 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.776930 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:00 crc kubenswrapper[4706]: E1127 07:09:00.776969 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.935638 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.935681 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.935692 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.935703 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.935715 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.935727 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.937122 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cb8d10b-fddc-42e4-9181-6366603ea340" containerID="b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd" exitCode=0 Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.937148 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerDied","Data":"b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd"} Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.949109 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.958921 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.971538 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.985816 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:00 crc kubenswrapper[4706]: I1127 07:09:00.999155 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:00Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.011299 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.022774 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.033681 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.045211 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.056387 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.070200 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.086458 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.097973 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.289802 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2hssd"] Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.290179 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.291686 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.292365 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.293533 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.294543 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.317366 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.326527 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.340015 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.351674 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.363351 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.374810 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.388242 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.390148 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n5t6\" (UniqueName: \"kubernetes.io/projected/026dfb3b-77c4-48bf-9402-a6fedc63f55a-kube-api-access-2n5t6\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.390186 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/026dfb3b-77c4-48bf-9402-a6fedc63f55a-host\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.390215 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/026dfb3b-77c4-48bf-9402-a6fedc63f55a-serviceca\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.397943 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.407086 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.417433 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.426583 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.437635 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.453077 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.464892 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.491404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/026dfb3b-77c4-48bf-9402-a6fedc63f55a-serviceca\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.491470 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n5t6\" (UniqueName: \"kubernetes.io/projected/026dfb3b-77c4-48bf-9402-a6fedc63f55a-kube-api-access-2n5t6\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.491501 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/026dfb3b-77c4-48bf-9402-a6fedc63f55a-host\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.491565 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/026dfb3b-77c4-48bf-9402-a6fedc63f55a-host\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.492665 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/026dfb3b-77c4-48bf-9402-a6fedc63f55a-serviceca\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.510724 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n5t6\" (UniqueName: \"kubernetes.io/projected/026dfb3b-77c4-48bf-9402-a6fedc63f55a-kube-api-access-2n5t6\") pod \"node-ca-2hssd\" (UID: \"026dfb3b-77c4-48bf-9402-a6fedc63f55a\") " pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.603272 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2hssd" Nov 27 07:09:01 crc kubenswrapper[4706]: W1127 07:09:01.620109 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod026dfb3b_77c4_48bf_9402_a6fedc63f55a.slice/crio-f926aa634f87f277650689f41b2f3ab110b7c60366a73943a3ece7a0b8e0f666 WatchSource:0}: Error finding container f926aa634f87f277650689f41b2f3ab110b7c60366a73943a3ece7a0b8e0f666: Status 404 returned error can't find the container with id f926aa634f87f277650689f41b2f3ab110b7c60366a73943a3ece7a0b8e0f666 Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.942431 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cb8d10b-fddc-42e4-9181-6366603ea340" containerID="24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9" exitCode=0 Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.942597 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerDied","Data":"24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9"} Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.944773 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2hssd" event={"ID":"026dfb3b-77c4-48bf-9402-a6fedc63f55a","Type":"ContainerStarted","Data":"d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d"} Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.944869 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2hssd" event={"ID":"026dfb3b-77c4-48bf-9402-a6fedc63f55a","Type":"ContainerStarted","Data":"f926aa634f87f277650689f41b2f3ab110b7c60366a73943a3ece7a0b8e0f666"} Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.961243 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.974603 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:01 crc kubenswrapper[4706]: I1127 07:09:01.988371 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.000761 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:01Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.015334 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.027687 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.045249 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.069713 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.093601 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.111305 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.127493 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.147654 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.162096 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.178843 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.200648 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.211518 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.229559 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.243959 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.256470 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.272093 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.287956 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.318003 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.358369 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.398944 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.438199 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.481165 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.521517 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.559700 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.775979 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:02 crc kubenswrapper[4706]: E1127 07:09:02.776111 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.776344 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:02 crc kubenswrapper[4706]: E1127 07:09:02.776651 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.776686 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:02 crc kubenswrapper[4706]: E1127 07:09:02.776857 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.951729 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.953988 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cb8d10b-fddc-42e4-9181-6366603ea340" containerID="b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767" exitCode=0 Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.954468 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerDied","Data":"b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767"} Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.970917 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.986099 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:02 crc kubenswrapper[4706]: I1127 07:09:02.998899 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:02Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.013180 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.026027 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.045842 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.057000 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.071852 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.084602 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.097376 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.111335 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.123274 4706 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.125006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.125043 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.125053 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.125235 4706 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.125759 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.132478 4706 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.132822 4706 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.133985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.134033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.134045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.134065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.134076 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.140626 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: E1127 07:09:03.146667 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.153575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.153626 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.153635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.153652 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.153666 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.156551 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: E1127 07:09:03.168647 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.172215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.172268 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.172282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.172297 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.172309 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: E1127 07:09:03.184087 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.187069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.187133 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.187146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.187163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.187174 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: E1127 07:09:03.198836 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.203109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.203146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.203155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.203183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.203194 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: E1127 07:09:03.214705 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: E1127 07:09:03.214962 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.217703 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.217739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.217748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.217762 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.217771 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.321632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.321921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.321930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.321944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.321954 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.424822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.424880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.424903 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.424933 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.424952 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.527719 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.527795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.527804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.527825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.527834 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.629822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.629856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.629864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.629879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.629887 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.732548 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.732589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.732599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.732618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.732628 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.834886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.834926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.834937 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.834951 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.834960 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.938050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.938079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.938087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.938099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.938107 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:03Z","lastTransitionTime":"2025-11-27T07:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.959482 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cb8d10b-fddc-42e4-9181-6366603ea340" containerID="0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc" exitCode=0 Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.959521 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerDied","Data":"0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc"} Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.971118 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.981866 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:03 crc kubenswrapper[4706]: I1127 07:09:03.993651 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:03Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.009186 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.020941 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.033576 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.040535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.040571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.040582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.040596 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.040608 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.046031 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.055649 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.072524 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.086744 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.098131 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.114662 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.124680 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.139582 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.142835 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.142861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.142871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.142886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.142897 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.244757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.244787 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.244795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.244809 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.244817 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.347519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.347573 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.347587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.347602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.347611 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.423381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.423542 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.423587 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.423651 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.423736 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:09:12.423702558 +0000 UTC m=+36.313293358 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.423651 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.423773 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:12.423763749 +0000 UTC m=+36.313354559 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.423801 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:12.42378493 +0000 UTC m=+36.313375740 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.449498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.449531 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.449541 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.449557 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.449570 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.524796 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.524843 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.524995 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.525015 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.525027 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.525026 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.525059 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.525070 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.525086 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:12.525067266 +0000 UTC m=+36.414658076 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.525122 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:12.525103457 +0000 UTC m=+36.414694257 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.551722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.551757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.551765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.551780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.551792 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.655104 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.655145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.655153 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.655167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.655177 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.757802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.757849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.757859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.757875 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.757886 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.775763 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.775793 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.775843 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.775887 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.775961 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:04 crc kubenswrapper[4706]: E1127 07:09:04.776016 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.860850 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.861355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.861388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.861405 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.861415 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.963030 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.963099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.963112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.963127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.963138 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:04Z","lastTransitionTime":"2025-11-27T07:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.965468 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cb8d10b-fddc-42e4-9181-6366603ea340" containerID="9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7" exitCode=0 Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.965548 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerDied","Data":"9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.972393 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7"} Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.972773 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.972830 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.972968 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.981114 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:04 crc kubenswrapper[4706]: I1127 07:09:04.996470 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:04Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.007635 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.010778 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.010952 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.028690 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.043781 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.055730 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.065313 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.065348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.065360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.065375 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.065385 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.069145 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.080358 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.096424 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.109765 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.124121 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.142654 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.152766 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.163170 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.168485 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.168514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.168524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.168539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.168549 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.173756 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.185206 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.197490 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.209396 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.219157 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.228118 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.237597 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.246242 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.256646 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.267810 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.271055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.271083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.271092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.271108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.271117 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.278178 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.288149 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.302777 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.312806 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.372758 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.372791 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.372799 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.372814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.372824 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.475063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.475395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.475516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.475609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.475691 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.577767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.577994 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.578097 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.578177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.578264 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.680727 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.680774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.680785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.680800 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.680811 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.784001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.784037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.784049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.784064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.784075 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.886085 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.886118 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.886126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.886139 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.886149 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.977997 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" event={"ID":"8cb8d10b-fddc-42e4-9181-6366603ea340","Type":"ContainerStarted","Data":"8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.988804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.988853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.988866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.988887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.988899 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:05Z","lastTransitionTime":"2025-11-27T07:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:05 crc kubenswrapper[4706]: I1127 07:09:05.991029 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.000189 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:05Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.010007 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.023437 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.035974 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.051717 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.065141 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.076323 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.090138 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.091635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.091897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.091913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.095284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.095353 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.106878 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.119092 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.145532 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.154184 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.166251 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.197135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.197182 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.197196 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.197249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.197266 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.300008 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.300044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.300052 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.300065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.300074 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.402594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.402643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.402656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.402675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.402691 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.504918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.504952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.504963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.504993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.505001 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.606856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.606889 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.606898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.606910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.606919 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.709985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.710032 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.710046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.710064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.710077 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.775996 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.776044 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.776105 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:06 crc kubenswrapper[4706]: E1127 07:09:06.776131 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:06 crc kubenswrapper[4706]: E1127 07:09:06.776207 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:06 crc kubenswrapper[4706]: E1127 07:09:06.776360 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.787664 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.800105 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.812875 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.812960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.812970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.812987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.812998 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.816634 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.829084 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.842772 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.854708 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.865720 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.881776 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.900110 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.912092 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.915429 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.915461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.915473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.915489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.915500 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:06Z","lastTransitionTime":"2025-11-27T07:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.927816 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.949612 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.960414 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:06 crc kubenswrapper[4706]: I1127 07:09:06.971597 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:06Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.018514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.018562 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.018574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.018592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.018604 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.121058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.121094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.121102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.121117 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.121125 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.223513 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.223591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.223602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.223617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.223626 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.325993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.326027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.326038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.326054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.326064 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.428189 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.428248 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.428261 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.428277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.428291 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.531091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.531402 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.531415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.531432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.531446 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.634707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.634748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.634759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.634775 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.634785 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.737521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.737560 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.737571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.737586 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.737597 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.840197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.840252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.840263 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.840280 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.840291 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.942438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.942532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.942546 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.942560 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.942569 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:07Z","lastTransitionTime":"2025-11-27T07:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.986376 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/0.log" Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.989028 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7" exitCode=1 Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.989075 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7"} Nov 27 07:09:07 crc kubenswrapper[4706]: I1127 07:09:07.989698 4706 scope.go:117] "RemoveContainer" containerID="43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.003333 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.016276 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.027896 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.035790 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.045475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.045516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.045530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.045553 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.045566 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.047355 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.058140 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.067713 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.087732 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:07Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626520 6011 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626603 6011 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626701 6011 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626761 6011 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626728 6011 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626809 6011 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.627436 6011 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:07.627455 6011 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:07.627467 6011 factory.go:656] Stopping watch factory\\\\nI1127 07:09:07.627478 6011 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:07.627496 6011 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.096984 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.107805 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.118137 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.133115 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.146610 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.148333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.148366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.148378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.148399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.148411 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.161000 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:08Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.250611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.250658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.250670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.250688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.250700 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.353176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.353216 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.353240 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.353254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.353263 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.455673 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.455719 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.455732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.455752 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.455764 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.557602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.557638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.557648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.557664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.557675 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.659679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.659717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.659726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.659743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.659757 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.762328 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.762367 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.762376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.762392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.762401 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.776772 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.776787 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.776809 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:08 crc kubenswrapper[4706]: E1127 07:09:08.776883 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:08 crc kubenswrapper[4706]: E1127 07:09:08.776962 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:08 crc kubenswrapper[4706]: E1127 07:09:08.777051 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.864621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.864657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.864665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.864678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.864688 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.966141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.966193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.966202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.966232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.966245 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:08Z","lastTransitionTime":"2025-11-27T07:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.996302 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/1.log" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.996913 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/0.log" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.999147 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b" exitCode=1 Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.999181 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b"} Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.999247 4706 scope.go:117] "RemoveContainer" containerID="43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7" Nov 27 07:09:08 crc kubenswrapper[4706]: I1127 07:09:08.999957 4706 scope.go:117] "RemoveContainer" containerID="e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b" Nov 27 07:09:09 crc kubenswrapper[4706]: E1127 07:09:09.000262 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.013852 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.023808 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.035663 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.046989 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.058624 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.068894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.068977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.068990 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.069006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.069021 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.075318 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:07Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626520 6011 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626603 6011 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626701 6011 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626761 6011 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626728 6011 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626809 6011 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.627436 6011 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:07.627455 6011 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:07.627467 6011 factory.go:656] Stopping watch factory\\\\nI1127 07:09:07.627478 6011 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:07.627496 6011 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.085915 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.097742 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.108616 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.120004 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.136342 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.148937 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.157870 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.167047 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:09Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.171842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.171860 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.171869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.171882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.171890 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.274533 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.274577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.274585 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.274604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.274615 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.376843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.376896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.376910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.376930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.376945 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.479999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.480041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.480051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.480067 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.480076 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.582198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.582283 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.582304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.582329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.582346 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.684738 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.684797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.684813 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.684833 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.684848 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.786891 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.786948 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.786964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.786987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.787003 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.889144 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.889202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.889241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.889266 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.889283 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.992606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.992675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.992699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.992726 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:09 crc kubenswrapper[4706]: I1127 07:09:09.992747 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:09Z","lastTransitionTime":"2025-11-27T07:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.007710 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/1.log" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.073037 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s"] Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.073688 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.076840 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.076907 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.090076 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.095304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.095343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.095355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.095371 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.095384 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.102355 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.117351 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.135440 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.148713 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.167043 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:07Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626520 6011 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626603 6011 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626701 6011 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626761 6011 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626728 6011 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626809 6011 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.627436 6011 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:07.627455 6011 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:07.627467 6011 factory.go:656] Stopping watch factory\\\\nI1127 07:09:07.627478 6011 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:07.627496 6011 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.176557 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.184194 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2mpz\" (UniqueName: \"kubernetes.io/projected/f6071ddb-03d9-4187-9e75-1abddee42d27-kube-api-access-q2mpz\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.184406 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f6071ddb-03d9-4187-9e75-1abddee42d27-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.184526 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f6071ddb-03d9-4187-9e75-1abddee42d27-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.184561 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f6071ddb-03d9-4187-9e75-1abddee42d27-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.188028 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.198124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.198162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.198175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.198189 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.198198 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.202751 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.215049 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.226281 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.239912 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.257099 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.269270 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.280648 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.285034 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f6071ddb-03d9-4187-9e75-1abddee42d27-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.285072 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f6071ddb-03d9-4187-9e75-1abddee42d27-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.285094 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2mpz\" (UniqueName: \"kubernetes.io/projected/f6071ddb-03d9-4187-9e75-1abddee42d27-kube-api-access-q2mpz\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.285128 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f6071ddb-03d9-4187-9e75-1abddee42d27-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.286063 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f6071ddb-03d9-4187-9e75-1abddee42d27-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.286294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f6071ddb-03d9-4187-9e75-1abddee42d27-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.291055 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f6071ddb-03d9-4187-9e75-1abddee42d27-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.300192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.300250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.300260 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.300276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.300286 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.311982 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2mpz\" (UniqueName: \"kubernetes.io/projected/f6071ddb-03d9-4187-9e75-1abddee42d27-kube-api-access-q2mpz\") pod \"ovnkube-control-plane-749d76644c-nf57s\" (UID: \"f6071ddb-03d9-4187-9e75-1abddee42d27\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.386439 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.401906 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.401947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.401960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.401977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.401988 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.506694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.507053 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.507065 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.507082 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.507093 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.609527 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.609565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.609574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.609589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.609598 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.711737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.711772 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.711781 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.711870 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.711880 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.754707 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.771820 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.775932 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.775946 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.775995 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:10 crc kubenswrapper[4706]: E1127 07:09:10.776067 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:10 crc kubenswrapper[4706]: E1127 07:09:10.776139 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:10 crc kubenswrapper[4706]: E1127 07:09:10.776258 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.788768 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.803477 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.814371 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.814418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.814430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.814449 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.814461 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.820418 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.838051 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.853595 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.864485 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.876183 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.885365 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.893625 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.902838 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.913827 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.916385 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.916418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.916427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.916441 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.916449 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:10Z","lastTransitionTime":"2025-11-27T07:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.924274 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.933847 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:10 crc kubenswrapper[4706]: I1127 07:09:10.949004 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:07Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626520 6011 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626603 6011 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626701 6011 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626761 6011 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626728 6011 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626809 6011 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.627436 6011 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:07.627455 6011 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:07.627467 6011 factory.go:656] Stopping watch factory\\\\nI1127 07:09:07.627478 6011 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:07.627496 6011 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:10Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.014815 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" event={"ID":"f6071ddb-03d9-4187-9e75-1abddee42d27","Type":"ContainerStarted","Data":"3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.014863 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" event={"ID":"f6071ddb-03d9-4187-9e75-1abddee42d27","Type":"ContainerStarted","Data":"b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.014874 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" event={"ID":"f6071ddb-03d9-4187-9e75-1abddee42d27","Type":"ContainerStarted","Data":"6e2a7fc42d6281605f97d05964814d9fe32bf98db83a0dfd48268bfdff1b3286"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.018256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.018300 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.018312 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.018330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.018341 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.027868 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.038083 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.048206 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.058831 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.073345 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.086126 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.097207 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.110271 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.118749 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.120081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.120118 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.120130 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.120145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.120155 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.134291 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.147701 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.162675 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.197106 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:07Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626520 6011 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626603 6011 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626701 6011 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626761 6011 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626728 6011 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626809 6011 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.627436 6011 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:07.627455 6011 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:07.627467 6011 factory.go:656] Stopping watch factory\\\\nI1127 07:09:07.627478 6011 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:07.627496 6011 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.212277 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.222054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.222082 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.222090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.222102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.222112 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.223256 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.323874 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.323917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.323926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.323938 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.323947 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.426398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.426634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.426643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.426657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.426665 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.527981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.528014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.528023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.528036 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.528045 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.630479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.630534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.630552 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.630575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.630590 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.732307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.732346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.732396 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.732415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.732427 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.834736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.834768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.834779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.834793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.834803 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.908628 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-zxp2h"] Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.909451 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:11 crc kubenswrapper[4706]: E1127 07:09:11.909573 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.927423 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.937271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.937314 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.937325 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.937344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.937356 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:11Z","lastTransitionTime":"2025-11-27T07:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.939287 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.954525 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.967715 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:11 crc kubenswrapper[4706]: I1127 07:09:11.985335 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.001993 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:11Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.002864 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.003033 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxrlx\" (UniqueName: \"kubernetes.io/projected/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-kube-api-access-vxrlx\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.015395 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.028748 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.039303 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.039343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.039352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.039367 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.039378 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.043753 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.052787 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.086409 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.104043 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxrlx\" (UniqueName: \"kubernetes.io/projected/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-kube-api-access-vxrlx\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.104118 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.104257 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.104337 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:09:12.604317989 +0000 UTC m=+36.493908799 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.117580 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.131337 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.132848 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxrlx\" (UniqueName: \"kubernetes.io/projected/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-kube-api-access-vxrlx\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.141211 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.141305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.141314 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.141328 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.141337 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.148362 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:07Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626520 6011 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626603 6011 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626701 6011 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626761 6011 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626728 6011 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626809 6011 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.627436 6011 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:07.627455 6011 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:07.627467 6011 factory.go:656] Stopping watch factory\\\\nI1127 07:09:07.627478 6011 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:07.627496 6011 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.157136 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.171019 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:12Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.243106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.243149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.243161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.243177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.243189 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.344900 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.344934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.344945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.344960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.344971 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.447643 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.447670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.447680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.447693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.447702 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.508837 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.508959 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.509002 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:09:28.508965637 +0000 UTC m=+52.398556487 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.509059 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.509057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.509094 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:28.509085931 +0000 UTC m=+52.398676741 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.509182 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.509288 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:28.509264986 +0000 UTC m=+52.398855866 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.550427 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.550468 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.550480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.550498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.550511 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.610728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.610792 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.610831 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.610898 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.610963 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:09:13.610944185 +0000 UTC m=+37.500534995 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.610964 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.610999 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.611018 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.611043 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.611068 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:28.611055848 +0000 UTC m=+52.500646668 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.611071 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.611096 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.611167 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:09:28.61114196 +0000 UTC m=+52.500732800 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.653319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.653376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.653392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.653415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.653431 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.755876 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.755909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.755917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.755949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.755959 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.776487 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.776529 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.776783 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.776795 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.776848 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:12 crc kubenswrapper[4706]: E1127 07:09:12.776949 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.858366 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.858454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.858471 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.858500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.858517 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.961320 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.961387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.961407 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.961429 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:12 crc kubenswrapper[4706]: I1127 07:09:12.961446 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:12Z","lastTransitionTime":"2025-11-27T07:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.064269 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.064335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.064360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.064389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.064408 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.167274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.168381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.168436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.168470 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.168494 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.272038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.272095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.272112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.272137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.272155 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.375697 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.375761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.375785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.375816 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.375914 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.419346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.419402 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.419419 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.419443 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.419462 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.440800 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:13Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.445594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.446021 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.446193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.446400 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.446558 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.467302 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:13Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.472261 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.472317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.472336 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.472359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.472375 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.486123 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:13Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.489936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.489979 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.489991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.490008 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.490020 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.503040 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:13Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.507173 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.507244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.507257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.507273 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.507287 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.520479 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:13Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.520597 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.522047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.522069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.522078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.522099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.522210 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.620882 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.621048 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.621342 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:09:15.621322727 +0000 UTC m=+39.510913537 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.625467 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.625498 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.625513 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.625528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.625538 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.728291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.728341 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.728352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.728372 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.728384 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.775711 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:13 crc kubenswrapper[4706]: E1127 07:09:13.775843 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.830591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.830819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.830897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.831009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.831077 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.933438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.933483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.933495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.933512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:13 crc kubenswrapper[4706]: I1127 07:09:13.933522 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:13Z","lastTransitionTime":"2025-11-27T07:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.035437 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.035470 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.035478 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.035491 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.035500 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.138006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.138055 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.138068 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.138086 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.138098 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.240236 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.240276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.240289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.240307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.240318 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.343409 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.343453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.343463 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.343480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.343494 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.445335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.445370 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.445378 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.445392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.445402 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.547399 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.547431 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.547439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.547451 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.547460 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.650344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.650669 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.650822 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.650958 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.651117 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.754099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.754136 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.754145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.754164 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.754175 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.775800 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.775822 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.775800 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:14 crc kubenswrapper[4706]: E1127 07:09:14.775919 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:14 crc kubenswrapper[4706]: E1127 07:09:14.776014 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:14 crc kubenswrapper[4706]: E1127 07:09:14.776179 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.857028 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.857094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.857110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.857131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.857147 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.959353 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.959388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.959398 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.959413 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:14 crc kubenswrapper[4706]: I1127 07:09:14.959426 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:14Z","lastTransitionTime":"2025-11-27T07:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.061678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.061710 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.061718 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.061732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.061744 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.164418 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.164447 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.164457 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.164471 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.164480 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.267979 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.268076 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.268096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.268121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.268139 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.370114 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.370167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.370181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.370200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.370214 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.472355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.472417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.472433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.472456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.472471 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.575358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.575414 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.575431 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.575451 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.575466 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.642612 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:15 crc kubenswrapper[4706]: E1127 07:09:15.642866 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:15 crc kubenswrapper[4706]: E1127 07:09:15.642959 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:09:19.642929679 +0000 UTC m=+43.532520529 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.678093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.678143 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.678159 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.678178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.678195 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.776610 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:15 crc kubenswrapper[4706]: E1127 07:09:15.776822 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.780790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.780937 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.781036 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.781101 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.781154 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.884395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.884486 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.884506 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.885034 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.885099 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.989255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.989549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.989638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.989725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:15 crc kubenswrapper[4706]: I1127 07:09:15.989823 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:15Z","lastTransitionTime":"2025-11-27T07:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.093003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.093054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.093070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.093094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.093109 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.196792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.196834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.196844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.196861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.196873 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.299835 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.299895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.299908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.299932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.299948 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.402330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.402379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.402389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.402402 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.402412 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.505567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.505634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.505644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.505660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.505669 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.608093 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.608170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.608194 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.608270 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.608296 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.711257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.711290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.711299 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.711314 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.711324 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.776847 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:16 crc kubenswrapper[4706]: E1127 07:09:16.777032 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.777575 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:16 crc kubenswrapper[4706]: E1127 07:09:16.777743 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.777802 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:16 crc kubenswrapper[4706]: E1127 07:09:16.778106 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.797897 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.814702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.814756 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.814764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.814779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.814809 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.820537 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.841803 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.867871 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.892039 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.913323 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.917620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.917665 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.917673 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.917690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.917702 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:16Z","lastTransitionTime":"2025-11-27T07:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.928624 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.940659 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.956398 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.974328 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:16 crc kubenswrapper[4706]: I1127 07:09:16.989914 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:16Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.010829 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43759a80482ae4a459043a7e99a8fb402bdc1b2058fc1b0061c3b1a5739adbb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:07Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626520 6011 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626603 6011 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626701 6011 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626761 6011 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.626728 6011 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 07:09:07.626809 6011 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 07:09:07.627436 6011 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:07.627455 6011 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:07.627467 6011 factory.go:656] Stopping watch factory\\\\nI1127 07:09:07.627478 6011 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:07.627496 6011 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:17Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.020115 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.020179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.020193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.020214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.020244 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:17Z","lastTransitionTime":"2025-11-27T07:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.022564 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:17Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.034319 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:17Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.046626 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:17Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.056715 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:17Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.122663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.122708 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.122717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.122735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.122745 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:17Z","lastTransitionTime":"2025-11-27T07:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.224981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.225206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.225333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.225397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.225460 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:17Z","lastTransitionTime":"2025-11-27T07:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.328964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.329027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.329050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.329080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:17 crc kubenswrapper[4706]: I1127 07:09:17.329102 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:17Z","lastTransitionTime":"2025-11-27T07:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.026355 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:18 crc kubenswrapper[4706]: E1127 07:09:18.026492 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.030892 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.030922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.030931 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.030946 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.030957 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.133779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.133815 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.133826 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.133842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.133854 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.236535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.236662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.236681 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.236705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.236722 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.339607 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.339660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.339676 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.339699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.339715 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.442782 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.442854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.442871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.442894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.442910 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.546629 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.546712 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.546746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.546776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.546800 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.650820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.650895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.650915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.650944 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.650967 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.755653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.755723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.755744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.755769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.755784 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.776213 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.776340 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:18 crc kubenswrapper[4706]: E1127 07:09:18.776462 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:18 crc kubenswrapper[4706]: E1127 07:09:18.776538 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.776255 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:18 crc kubenswrapper[4706]: E1127 07:09:18.776695 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.859975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.860047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.860071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.860107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.860121 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.964250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.964336 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.964361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.964392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:18 crc kubenswrapper[4706]: I1127 07:09:18.964417 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:18Z","lastTransitionTime":"2025-11-27T07:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.071997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.072069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.072092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.072139 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.072193 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.175543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.175599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.175613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.175635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.175653 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.278819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.278897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.278917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.278949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.278972 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.381658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.381720 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.381733 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.381759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.381775 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.484574 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.484633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.484650 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.484674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.484689 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.589210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.589332 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.589352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.589383 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.589405 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.644186 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:19 crc kubenswrapper[4706]: E1127 07:09:19.644525 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:19 crc kubenswrapper[4706]: E1127 07:09:19.644730 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:09:27.644683485 +0000 UTC m=+51.534274455 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.692897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.692947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.692962 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.692984 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.693000 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.776643 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:19 crc kubenswrapper[4706]: E1127 07:09:19.776810 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.796350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.796415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.796439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.796470 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.796496 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.899135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.899183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.899200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.899269 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:19 crc kubenswrapper[4706]: I1127 07:09:19.899295 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:19Z","lastTransitionTime":"2025-11-27T07:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.002057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.002122 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.002140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.002163 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.002183 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.105430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.105464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.105475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.105492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.105502 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.208387 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.208463 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.208488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.208522 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.208546 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.310461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.310501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.310511 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.310526 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.310536 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.413893 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.413968 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.413993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.414029 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.414054 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.517517 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.517570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.517589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.517615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.517632 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.621284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.621356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.621379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.621409 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.621432 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.724349 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.724406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.724417 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.724436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.724448 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.776753 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:20 crc kubenswrapper[4706]: E1127 07:09:20.776893 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.776704 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:20 crc kubenswrapper[4706]: E1127 07:09:20.776976 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.776757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:20 crc kubenswrapper[4706]: E1127 07:09:20.777043 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.827492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.827520 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.827528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.827543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.827552 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.929762 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.929790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.929798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.929872 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:20 crc kubenswrapper[4706]: I1127 07:09:20.929884 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:20Z","lastTransitionTime":"2025-11-27T07:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.032679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.032780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.032808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.032843 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.032867 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.137142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.137202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.137263 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.137313 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.137351 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.239637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.239694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.239713 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.239739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.239756 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.342152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.342195 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.342206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.342253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.342271 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.444876 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.444940 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.444963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.444993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.445016 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.548709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.548763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.548779 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.548805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.548822 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.651747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.651786 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.651794 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.651808 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.651821 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.755010 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.755087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.755107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.755132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.755149 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.775908 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:21 crc kubenswrapper[4706]: E1127 07:09:21.776136 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.857959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.857996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.858005 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.858020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.858030 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.961150 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.961207 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.961243 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.961264 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:21 crc kubenswrapper[4706]: I1127 07:09:21.961278 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:21Z","lastTransitionTime":"2025-11-27T07:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.065089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.065137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.065155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.065178 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.065194 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.168472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.168529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.168547 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.168571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.168592 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.271837 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.271883 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.271894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.271911 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.271923 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.374315 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.374360 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.374371 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.374389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.374403 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.476437 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.476503 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.476515 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.476534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.476548 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.579958 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.580028 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.580038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.580061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.580075 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.682277 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.682313 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.682323 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.682341 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.682353 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.775913 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.775956 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:22 crc kubenswrapper[4706]: E1127 07:09:22.776076 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.776105 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:22 crc kubenswrapper[4706]: E1127 07:09:22.776298 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:22 crc kubenswrapper[4706]: E1127 07:09:22.776356 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.784793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.784834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.784846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.784861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.784873 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.887615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.887649 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.887657 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.887670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.887679 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.989865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.989899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.989908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.989921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:22 crc kubenswrapper[4706]: I1127 07:09:22.989930 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:22Z","lastTransitionTime":"2025-11-27T07:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.093302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.093393 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.093411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.093438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.093458 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.196913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.196982 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.197004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.197033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.197053 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.300384 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.300455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.300478 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.300507 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.300528 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.402496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.402556 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.402572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.402597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.402613 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.505138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.505202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.505254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.505291 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.505314 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.607534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.607569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.607581 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.607595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.607606 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.710284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.710318 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.710329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.710346 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.710359 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.776192 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:23 crc kubenswrapper[4706]: E1127 07:09:23.776406 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.777652 4706 scope.go:117] "RemoveContainer" containerID="e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.790540 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.801016 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.813853 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.814061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.814108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.814146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.814166 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.814178 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.826710 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.839312 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.855424 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.865187 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.876511 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.886708 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.897864 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.909987 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.917112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.917145 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.917154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.917170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.917180 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.920782 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.920828 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.920838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.920854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.920865 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.927794 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: E1127 07:09:23.931694 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.934614 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.934658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.934668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.934683 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.934695 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.940484 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: E1127 07:09:23.947678 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.950690 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.951477 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.951509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.951518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.951531 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.951539 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.963382 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: E1127 07:09:23.965153 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.968215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.968254 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.968264 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.968290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.968301 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.972400 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: E1127 07:09:23.980562 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.983198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.983248 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.983260 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.983274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:23 crc kubenswrapper[4706]: I1127 07:09:23.983282 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:23Z","lastTransitionTime":"2025-11-27T07:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:23 crc kubenswrapper[4706]: E1127 07:09:23.993009 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:23Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:23 crc kubenswrapper[4706]: E1127 07:09:23.993123 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.019549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.019589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.019598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.019613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.019624 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.062184 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/1.log" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.064669 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.065067 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.076298 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.086084 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.096482 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.108114 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.120497 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.122157 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.122190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.122202 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.122238 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.122252 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.138135 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.154202 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.168418 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.181991 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.193899 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.216432 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.224735 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.224790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.224801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.224819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.224830 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.236189 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.257603 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.277370 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.288054 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.298514 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.327634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.327691 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.327701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.327721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.327733 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.346366 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.354918 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.362108 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.376977 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.389891 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.402056 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.414550 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.423879 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.430292 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.430336 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.430376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.430394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.430409 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.437213 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.449034 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.462759 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.483157 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.493719 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.506845 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.522174 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.532550 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.532582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.532591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.532609 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.532620 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.536322 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.549124 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.570585 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:24Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.634523 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.634564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.634576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.634595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.634607 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.737066 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.737115 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.737125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.737140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.737148 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.775722 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:24 crc kubenswrapper[4706]: E1127 07:09:24.775845 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.775733 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.775913 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:24 crc kubenswrapper[4706]: E1127 07:09:24.776122 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:24 crc kubenswrapper[4706]: E1127 07:09:24.776212 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.839298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.839379 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.839403 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.839438 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.839464 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.942381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.942433 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.942446 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.942465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:24 crc kubenswrapper[4706]: I1127 07:09:24.942478 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:24Z","lastTransitionTime":"2025-11-27T07:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.045107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.045155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.045170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.045190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.045205 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.068715 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/2.log" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.069690 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/1.log" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.072699 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628" exitCode=1 Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.072838 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.072917 4706 scope.go:117] "RemoveContainer" containerID="e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.073748 4706 scope.go:117] "RemoveContainer" containerID="9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628" Nov 27 07:09:25 crc kubenswrapper[4706]: E1127 07:09:25.073992 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.091088 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.105657 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.119414 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.135757 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.148768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.148807 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.148818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.148835 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.148844 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.157041 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.169544 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.185941 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.197121 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.207608 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.219805 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.231725 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.241121 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.251448 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.251494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.251506 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.251525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.251537 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.261247 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f89b3233fe2b77a1d5b6c7533579163acedd05ef0b467ab1ce3db7144b0a5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:08Z\\\",\\\"message\\\":\\\"rvices.LB{}\\\\nI1127 07:09:08.830171 6138 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 07:09:08.830193 6138 services_controller.go:454] Service openshift-apiserver/api for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 07:09:08.830196 6138 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.270826 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.282478 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.295302 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.308440 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:25Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.354465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.354544 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.354570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.354599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.354623 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.457117 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.457153 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.457165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.457181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.457192 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.560214 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.560283 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.560293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.560310 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.560321 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.663932 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.664001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.664019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.664063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.664083 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.766516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.766592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.766615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.766651 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.766674 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.776126 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:25 crc kubenswrapper[4706]: E1127 07:09:25.776358 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.870422 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.870506 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.870519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.870535 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.870546 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.973547 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.973587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.973597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.973610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:25 crc kubenswrapper[4706]: I1127 07:09:25.973619 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:25Z","lastTransitionTime":"2025-11-27T07:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.075412 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.075454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.075464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.075481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.075492 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.078851 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/2.log" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.082972 4706 scope.go:117] "RemoveContainer" containerID="9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628" Nov 27 07:09:26 crc kubenswrapper[4706]: E1127 07:09:26.083278 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.097357 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.111674 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.134949 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.155896 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.173525 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.178193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.178245 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.178255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.178272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.178282 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.192439 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.202948 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.216538 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.228413 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.238713 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.248250 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.263205 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.275318 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.280578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.280748 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.280828 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.280915 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.280995 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.286666 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.296443 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.306174 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.313537 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.383255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.383286 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.383297 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.383311 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.383322 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.485699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.486049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.486187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.486359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.486522 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.589081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.589122 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.589131 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.589144 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.589154 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.691436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.691792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.691861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.691947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.692026 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.776128 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.776128 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.776264 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:26 crc kubenswrapper[4706]: E1127 07:09:26.776399 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:26 crc kubenswrapper[4706]: E1127 07:09:26.776462 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:26 crc kubenswrapper[4706]: E1127 07:09:26.776693 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.790948 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.794664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.794910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.794975 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.795039 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.795106 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.805911 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.816858 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.832639 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.845872 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.861995 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.878641 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.890786 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.897162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.897200 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.897212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.897263 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.897276 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.901659 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.917796 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.929315 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.946297 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.959358 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.972143 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.993174 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:26Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.999664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.999722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.999737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.999759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:26 crc kubenswrapper[4706]: I1127 07:09:26.999774 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:26Z","lastTransitionTime":"2025-11-27T07:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.002387 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:27Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.012886 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:27Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.101818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.101856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.101866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.101880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.101889 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.203536 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.203575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.203587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.203604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.203619 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.305606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.305644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.305656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.305670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.305679 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.407479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.407513 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.407521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.407534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.407543 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.510475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.510502 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.510511 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.510524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.510533 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.613148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.613193 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.613206 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.613241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.613251 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.715959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.716115 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.716128 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.716146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.716160 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.734768 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:27 crc kubenswrapper[4706]: E1127 07:09:27.734918 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:27 crc kubenswrapper[4706]: E1127 07:09:27.735044 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:09:43.734967081 +0000 UTC m=+67.624557901 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.776509 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:27 crc kubenswrapper[4706]: E1127 07:09:27.776698 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.819138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.819191 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.819210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.819260 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.819277 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.921965 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.921991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.922003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.922015 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:27 crc kubenswrapper[4706]: I1127 07:09:27.922026 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:27Z","lastTransitionTime":"2025-11-27T07:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.025184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.025256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.025271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.025284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.025292 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.127661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.127696 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.127704 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.127717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.127725 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.232769 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.232836 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.232854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.232879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.232895 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.335405 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.335457 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.335472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.335493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.335509 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.438151 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.438288 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.438309 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.438699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.438927 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.541394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.541437 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.541445 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.541461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.541470 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.543918 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.544092 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.544139 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:10:00.544105833 +0000 UTC m=+84.433696683 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.544196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.544251 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.544308 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.544346 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:10:00.544326579 +0000 UTC m=+84.433917409 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.544614 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:10:00.54435815 +0000 UTC m=+84.433948970 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.644430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.644480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.644490 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.644514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.644523 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.644957 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.645011 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645198 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645257 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645268 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645319 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645343 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645276 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645419 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:10:00.645388249 +0000 UTC m=+84.534979099 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.645465 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:10:00.64544087 +0000 UTC m=+84.535031760 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.747666 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.747737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.747755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.747809 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.747830 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.776918 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.777092 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.777481 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.777569 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.777891 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:28 crc kubenswrapper[4706]: E1127 07:09:28.778024 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.850651 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.850690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.850699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.850714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.850724 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.953557 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.953630 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.953644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.953660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:28 crc kubenswrapper[4706]: I1127 07:09:28.953693 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:28Z","lastTransitionTime":"2025-11-27T07:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.056357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.056659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.056794 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.056917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.057034 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.159329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.159489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.159590 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.159685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.159770 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.261871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.261989 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.261999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.262014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.262023 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.364046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.364086 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.364094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.364109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.364120 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.467061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.467099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.467107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.467119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.467128 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.569659 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.569699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.569709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.569723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.569732 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.672268 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.672320 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.672329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.672342 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.672350 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.774755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.774829 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.774850 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.774887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.774909 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.776068 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:29 crc kubenswrapper[4706]: E1127 07:09:29.776287 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.878337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.878381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.878389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.878404 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.878416 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.981175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.981249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.981262 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.981279 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:29 crc kubenswrapper[4706]: I1127 07:09:29.981291 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:29Z","lastTransitionTime":"2025-11-27T07:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.083390 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.083449 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.083458 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.083475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.083484 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.186966 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.187002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.187012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.187026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.187035 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.289052 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.289113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.289127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.289147 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.289556 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.391834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.391898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.391908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.391924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.391934 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.494187 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.494247 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.494257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.494272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.494313 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.596667 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.596717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.596729 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.596743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.596752 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.699620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.699683 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.699695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.699709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.699718 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.776084 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.776135 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.776100 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:30 crc kubenswrapper[4706]: E1127 07:09:30.776287 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:30 crc kubenswrapper[4706]: E1127 07:09:30.776404 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:30 crc kubenswrapper[4706]: E1127 07:09:30.776465 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.801701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.801733 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.801741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.801753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.801762 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.904455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.904500 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.904545 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.904563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:30 crc kubenswrapper[4706]: I1127 07:09:30.904580 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:30Z","lastTransitionTime":"2025-11-27T07:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.006468 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.006507 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.006517 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.006530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.006540 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.108914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.108952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.108960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.108972 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.108981 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.211519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.211553 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.211561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.211577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.211586 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.314089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.314132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.314143 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.314159 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.314168 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.416707 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.416742 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.416751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.416765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.416773 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.519510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.519592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.519615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.519649 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.519675 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.622181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.622232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.622241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.622255 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.622264 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.725247 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.725296 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.725309 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.725326 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.725337 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.775757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:31 crc kubenswrapper[4706]: E1127 07:09:31.775934 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.827875 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.827942 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.827960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.827986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.828004 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.930232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.930271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.930283 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.930296 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:31 crc kubenswrapper[4706]: I1127 07:09:31.930305 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:31Z","lastTransitionTime":"2025-11-27T07:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.032819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.032859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.032867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.032886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.032896 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.135578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.135627 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.135636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.135649 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.135658 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.237924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.237972 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.237982 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.237997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.238008 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.340211 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.340268 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.340276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.340290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.340299 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.442778 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.442834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.442851 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.442874 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.442890 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.545355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.545388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.545397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.545409 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.545418 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.647583 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.647620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.647629 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.647644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.647655 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.750908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.750969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.750987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.751045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.751064 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.776517 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.776539 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:32 crc kubenswrapper[4706]: E1127 07:09:32.776734 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:32 crc kubenswrapper[4706]: E1127 07:09:32.776916 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.777036 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:32 crc kubenswrapper[4706]: E1127 07:09:32.777147 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.853766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.853824 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.853833 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.853875 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.853885 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.956442 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.956506 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.956527 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.956554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:32 crc kubenswrapper[4706]: I1127 07:09:32.956572 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:32Z","lastTransitionTime":"2025-11-27T07:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.059014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.059039 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.059064 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.059079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.059087 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.161705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.161753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.161767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.161799 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.161811 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.263899 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.263960 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.263977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.264002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.264019 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.366049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.366113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.366174 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.366203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.366254 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.468895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.468961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.468976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.468997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.469016 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.571964 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.572022 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.572039 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.572062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.572079 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.674121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.674161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.674170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.674186 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.674196 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.775812 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:33 crc kubenswrapper[4706]: E1127 07:09:33.776015 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.777638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.777702 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.777721 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.777745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.777761 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.880450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.880480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.880488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.880501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.880511 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.982958 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.983002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.983016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.983037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:33 crc kubenswrapper[4706]: I1127 07:09:33.983051 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:33Z","lastTransitionTime":"2025-11-27T07:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.085974 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.086033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.086048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.086118 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.086155 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.189278 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.189313 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.189322 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.189337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.189347 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.208442 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.208544 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.208569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.208598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.208617 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.230539 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:34Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.236147 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.236185 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.236194 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.236208 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.236265 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.255081 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:34Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.259904 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.259935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.259943 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.259961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.259970 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.283152 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:34Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.289339 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.289371 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.289380 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.289395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.289407 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.309207 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:34Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.314140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.314168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.314177 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.314190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.314199 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.331297 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:34Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.331405 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.333045 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.333107 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.333129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.333203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.333305 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.436739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.436774 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.436783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.436798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.436808 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.539575 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.539694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.539718 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.539745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.539772 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.641992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.642023 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.642031 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.642044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.642054 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.743914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.743969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.743986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.744009 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.744028 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.775771 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.775824 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.775872 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.775897 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.776118 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:34 crc kubenswrapper[4706]: E1127 07:09:34.776549 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.847041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.847105 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.847124 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.847149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.847168 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.950171 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.950239 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.950253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.950271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:34 crc kubenswrapper[4706]: I1127 07:09:34.950284 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:34Z","lastTransitionTime":"2025-11-27T07:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.052827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.052884 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.052896 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.052914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.052926 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.155981 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.156060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.156083 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.156109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.156131 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.258973 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.259016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.259027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.259058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.259069 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.361765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.361797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.361804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.361818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.361826 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.465004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.465049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.465062 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.465078 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.465087 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.567113 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.567180 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.567198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.567250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.567294 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.670810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.670852 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.670861 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.670876 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.670886 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.773644 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.773699 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.773711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.773732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.773745 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.776076 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:35 crc kubenswrapper[4706]: E1127 07:09:35.776383 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.876775 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.876869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.876888 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.876918 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.876942 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.981634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.981740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.981766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.981801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:35 crc kubenswrapper[4706]: I1127 07:09:35.981838 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:35Z","lastTransitionTime":"2025-11-27T07:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.084089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.084155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.084170 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.084188 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.084200 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.186395 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.186435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.186444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.186458 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.186466 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.288752 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.288800 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.288812 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.288827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.288838 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.391213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.391305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.391322 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.391347 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.391365 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.493638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.493680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.493690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.493705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.493715 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.596630 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.596675 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.596685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.596701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.596711 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.698954 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.699279 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.699364 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.699444 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.699516 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.776116 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.776162 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.776116 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:36 crc kubenswrapper[4706]: E1127 07:09:36.776266 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:36 crc kubenswrapper[4706]: E1127 07:09:36.776312 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:36 crc kubenswrapper[4706]: E1127 07:09:36.776370 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.789663 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.802247 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.802293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.802307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.802325 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.802340 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.808507 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.819667 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.830718 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.846310 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.858645 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.870284 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.883337 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.893271 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.904746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.904925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.905050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.905190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.905341 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:36Z","lastTransitionTime":"2025-11-27T07:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.906638 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.918941 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.931069 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.941160 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.952570 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.961985 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.982648 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:36 crc kubenswrapper[4706]: I1127 07:09:36.994866 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:36Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.007411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.007447 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.007458 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.007472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.007481 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.109397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.109434 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.109447 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.109464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.109475 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.211841 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.211873 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.211882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.211897 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.211906 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.314722 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.314830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.314855 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.314886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.314909 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.417406 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.417449 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.417457 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.417472 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.417484 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.519589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.519633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.519651 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.519671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.519688 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.622250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.622333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.622343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.622361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.622375 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.725936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.725985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.725998 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.726014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.726026 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.775788 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:37 crc kubenswrapper[4706]: E1127 07:09:37.775938 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.828825 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.828858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.828866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.828879 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.828887 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.931091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.931130 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.931140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.931169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:37 crc kubenswrapper[4706]: I1127 07:09:37.931178 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:37Z","lastTransitionTime":"2025-11-27T07:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.033777 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.033832 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.033848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.033866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.033880 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.135457 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.135496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.135504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.135516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.135525 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.238122 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.238205 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.238249 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.238281 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.238300 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.340503 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.340549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.340561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.340580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.340592 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.443947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.443995 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.444006 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.444028 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.444038 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.547165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.547275 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.547294 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.547317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.547371 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.651088 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.651137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.651149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.651169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.651181 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.754640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.754880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.754942 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.755011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.755075 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.776131 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.776817 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.776807 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:38 crc kubenswrapper[4706]: E1127 07:09:38.776986 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:38 crc kubenswrapper[4706]: E1127 07:09:38.777273 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:38 crc kubenswrapper[4706]: E1127 07:09:38.777335 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.858063 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.858141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.858161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.858192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.858212 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.961810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.961873 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.961891 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.961916 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:38 crc kubenswrapper[4706]: I1127 07:09:38.961935 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:38Z","lastTransitionTime":"2025-11-27T07:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.066312 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.066350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.066361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.066377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.066389 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.169203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.169252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.169260 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.169275 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.169284 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.272388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.272441 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.272454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.272475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.272486 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.374557 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.374600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.374613 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.374631 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.374643 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.477908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.477949 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.477961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.477977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.477989 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.580621 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.580688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.580711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.580741 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.580763 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.683137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.683185 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.683199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.683241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.683258 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.775994 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:39 crc kubenswrapper[4706]: E1127 07:09:39.776113 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.785071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.785099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.785110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.785125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.785135 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.887695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.887776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.887789 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.887806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.887815 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.990755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.990807 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.990818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.990836 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:39 crc kubenswrapper[4706]: I1127 07:09:39.990849 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:39Z","lastTransitionTime":"2025-11-27T07:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.092612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.092661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.092674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.092692 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.092705 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.195049 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.195081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.195090 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.195103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.195112 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.297070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.297422 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.297588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.297700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.297857 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.400130 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.400459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.400548 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.400661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.400765 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.503530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.503568 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.503580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.503593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.503603 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.606118 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.606166 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.606182 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.606212 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.606255 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.708834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.708877 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.708889 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.708907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.708919 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.775935 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.775916 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.776027 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:40 crc kubenswrapper[4706]: E1127 07:09:40.776425 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:40 crc kubenswrapper[4706]: E1127 07:09:40.776501 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:40 crc kubenswrapper[4706]: E1127 07:09:40.776552 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.776786 4706 scope.go:117] "RemoveContainer" containerID="9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628" Nov 27 07:09:40 crc kubenswrapper[4706]: E1127 07:09:40.776958 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.811293 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.811330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.811340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.811356 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.811369 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.913637 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.913682 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.913693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.913709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:40 crc kubenswrapper[4706]: I1127 07:09:40.913719 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:40Z","lastTransitionTime":"2025-11-27T07:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.015842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.015880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.015890 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.015907 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.015917 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.118586 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.118617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.118625 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.118638 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.118646 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.220161 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.220253 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.220270 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.220288 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.220299 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.322330 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.322368 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.322376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.322391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.322403 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.424893 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.424934 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.424945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.424963 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.424974 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.527046 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.527091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.527100 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.527115 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.527128 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.629987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.630033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.630043 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.630059 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.630070 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.732684 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.732739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.732750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.732768 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.732778 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.776399 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:41 crc kubenswrapper[4706]: E1127 07:09:41.776559 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.835149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.835190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.835201 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.835237 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.835250 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.937190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.937272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.937290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.937364 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:41 crc kubenswrapper[4706]: I1127 07:09:41.937383 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:41Z","lastTransitionTime":"2025-11-27T07:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.039715 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.039745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.039753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.039767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.039775 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.142234 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.142282 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.142292 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.142309 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.142320 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.244605 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.244654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.244663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.244679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.244689 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.346806 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.346850 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.346865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.346883 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.346896 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.448610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.448641 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.448649 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.448661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.448670 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.550589 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.550651 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.550663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.550679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.550690 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.652467 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.652518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.652528 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.652539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.652548 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.754884 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.754914 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.754924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.754936 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.754947 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.776357 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.776410 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:42 crc kubenswrapper[4706]: E1127 07:09:42.776516 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.776538 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:42 crc kubenswrapper[4706]: E1127 07:09:42.776663 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:42 crc kubenswrapper[4706]: E1127 07:09:42.776755 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.856954 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.856992 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.857001 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.857013 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.857022 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.959295 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.959335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.959347 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.959362 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:42 crc kubenswrapper[4706]: I1127 07:09:42.959371 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:42Z","lastTransitionTime":"2025-11-27T07:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.061525 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.061559 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.061567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.061580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.061589 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.163752 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.163792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.163800 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.163814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.163823 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.265522 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.265566 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.265576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.265591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.265600 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.368431 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.368468 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.368480 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.368497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.368508 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.470732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.470785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.470797 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.470818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.470829 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.572856 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.572910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.572919 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.572938 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.572948 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.675105 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.675144 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.675152 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.675169 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.675178 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.776197 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:43 crc kubenswrapper[4706]: E1127 07:09:43.776463 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.777611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.777656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.777668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.777685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.777695 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.817113 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:43 crc kubenswrapper[4706]: E1127 07:09:43.817289 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:43 crc kubenswrapper[4706]: E1127 07:09:43.817388 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:10:15.817364333 +0000 UTC m=+99.706955143 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.879860 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.879902 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.879911 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.879927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.879938 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.981941 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.981985 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.981997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.982012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:43 crc kubenswrapper[4706]: I1127 07:09:43.982024 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:43Z","lastTransitionTime":"2025-11-27T07:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.084592 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.084632 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.084640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.084654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.084663 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.186867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.186935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.186948 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.186966 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.186978 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.288811 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.288851 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.288863 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.288878 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.288891 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.390532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.390571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.390580 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.390593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.390603 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.492424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.492466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.492481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.492497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.492505 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.514842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.514894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.514909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.514952 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.514963 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.525410 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:44Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.529154 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.529199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.529210 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.529241 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.529253 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.540461 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:44Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.543970 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.544033 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.544044 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.544059 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.544087 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.555326 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:44Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.558410 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.558484 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.558495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.558526 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.558539 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.571857 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:44Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.576588 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.576625 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.576634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.576648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.576657 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.587683 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:44Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.587837 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.594132 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.594167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.594175 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.594190 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.594200 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.696529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.696590 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.696604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.696619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.696627 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.776002 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.776010 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.776128 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.776152 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.776204 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:44 crc kubenswrapper[4706]: E1127 07:09:44.776294 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.798969 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.798999 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.799011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.799025 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.799035 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.901541 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.901579 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.901591 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.901608 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:44 crc kubenswrapper[4706]: I1127 07:09:44.901619 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:44Z","lastTransitionTime":"2025-11-27T07:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.003619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.004165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.004298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.004411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.004497 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.106670 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.106714 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.106723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.106739 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.106748 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.208567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.208610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.208618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.208634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.208644 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.310533 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.310565 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.310573 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.310587 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.310597 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.413153 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.413197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.413208 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.413227 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.413252 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.515140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.515176 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.515184 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.515197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.515208 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.617232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.617276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.617285 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.617301 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.617312 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.719900 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.719945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.719956 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.719971 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.719980 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.776697 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:45 crc kubenswrapper[4706]: E1127 07:09:45.776835 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.822678 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.822715 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.822724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.822737 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.822748 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.925674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.925729 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.925742 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.925763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:45 crc kubenswrapper[4706]: I1127 07:09:45.925775 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:45Z","lastTransitionTime":"2025-11-27T07:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.029593 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.029635 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.029652 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.029674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.029694 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.132048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.132099 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.132117 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.132141 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.132157 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.139191 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/0.log" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.139274 4706 generic.go:334] "Generic (PLEG): container finished" podID="0e33550a-5308-4e22-9c81-8a12e97f44ef" containerID="f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7" exitCode=1 Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.139303 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerDied","Data":"f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.139911 4706 scope.go:117] "RemoveContainer" containerID="f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.155522 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.170386 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.180348 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.191274 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.202644 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.220759 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:45Z\\\",\\\"message\\\":\\\"2025-11-27T07:08:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5\\\\n2025-11-27T07:08:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5 to /host/opt/cni/bin/\\\\n2025-11-27T07:09:00Z [verbose] multus-daemon started\\\\n2025-11-27T07:09:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T07:09:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.235415 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.235537 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.235554 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.235572 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.235585 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.237157 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.250231 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.259401 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.268626 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.276716 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.288529 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.301040 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.311211 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.327416 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.336665 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.337805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.337834 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.337846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.337865 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.337876 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.348292 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.439488 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.439516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.439524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.439539 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.439548 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.541728 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.541766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.541798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.541814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.541826 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.644162 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.644244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.644265 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.644287 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.644301 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.747198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.747230 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.747240 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.747271 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.747280 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.776289 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.776348 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:46 crc kubenswrapper[4706]: E1127 07:09:46.776381 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:46 crc kubenswrapper[4706]: E1127 07:09:46.776467 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.776564 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:46 crc kubenswrapper[4706]: E1127 07:09:46.776754 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.790374 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.800888 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.810528 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.820668 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.830846 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.841918 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:45Z\\\",\\\"message\\\":\\\"2025-11-27T07:08:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5\\\\n2025-11-27T07:08:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5 to /host/opt/cni/bin/\\\\n2025-11-27T07:09:00Z [verbose] multus-daemon started\\\\n2025-11-27T07:09:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T07:09:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.849290 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.849341 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.849355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.849374 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.849387 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.855626 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.866412 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.903179 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.922636 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.936071 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.945532 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.952129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.952167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.952179 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.952196 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.952205 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:46Z","lastTransitionTime":"2025-11-27T07:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.957795 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.967490 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.977126 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:46 crc kubenswrapper[4706]: I1127 07:09:46.992834 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.001500 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.054611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.054653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.054662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.054677 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.054688 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.143778 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/0.log" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.143835 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerStarted","Data":"5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.156799 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.156835 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.156848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.156866 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.156881 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.164071 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.172455 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.181508 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.193500 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.203854 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.215466 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.226107 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.235827 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.246481 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.258698 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.258742 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.258755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.258772 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.258785 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.264807 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.274643 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.285038 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.295535 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:45Z\\\",\\\"message\\\":\\\"2025-11-27T07:08:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5\\\\n2025-11-27T07:08:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5 to /host/opt/cni/bin/\\\\n2025-11-27T07:09:00Z [verbose] multus-daemon started\\\\n2025-11-27T07:09:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T07:09:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.305629 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.319880 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.329548 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.340217 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.361070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.361098 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.361106 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.361119 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.361128 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.463344 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.463373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.463381 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.463393 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.463403 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.565321 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.565361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.565372 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.565391 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.565402 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.667319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.667353 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.667361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.667376 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.667384 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.769997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.770042 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.770054 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.770071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.770082 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.776238 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:47 crc kubenswrapper[4706]: E1127 07:09:47.776340 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.872445 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.872487 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.872497 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.872512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.872522 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.974801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.974848 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.974862 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.974880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:47 crc kubenswrapper[4706]: I1127 07:09:47.974892 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:47Z","lastTransitionTime":"2025-11-27T07:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.077156 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.077203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.077232 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.077250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.077261 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.179600 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.179640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.179648 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.179660 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.179669 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.281646 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.281687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.281695 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.281713 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.281723 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.384668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.384731 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.384745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.384765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.384781 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.486429 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.486464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.486475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.486489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.486499 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.587842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.587868 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.587886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.587901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.587911 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.689878 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.689909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.689927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.689945 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.689956 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.776634 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:48 crc kubenswrapper[4706]: E1127 07:09:48.776746 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.776759 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:48 crc kubenswrapper[4706]: E1127 07:09:48.776830 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.776634 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:48 crc kubenswrapper[4706]: E1127 07:09:48.776894 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.791571 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.791602 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.791612 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.791625 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.791636 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.894026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.894085 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.894095 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.894109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.894120 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.996390 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.996432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.996441 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.996460 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:48 crc kubenswrapper[4706]: I1127 07:09:48.996471 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:48Z","lastTransitionTime":"2025-11-27T07:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.098982 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.099016 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.099024 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.099040 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.099050 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.201037 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.201071 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.201080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.201092 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.201101 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.302948 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.302984 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.302991 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.303003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.303013 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.405564 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.405598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.405606 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.405618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.405626 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.508229 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.508287 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.508295 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.508311 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.508320 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.611138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.611185 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.611194 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.611208 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.611235 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.713392 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.713436 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.713445 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.713459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.713469 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.775965 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:49 crc kubenswrapper[4706]: E1127 07:09:49.776122 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.815276 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.815309 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.815319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.815331 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.815341 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.917653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.917680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.917687 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.917700 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:49 crc kubenswrapper[4706]: I1127 07:09:49.917708 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:49Z","lastTransitionTime":"2025-11-27T07:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.019501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.019540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.019549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.019563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.019573 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.121266 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.121311 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.121329 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.121345 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.121356 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.223747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.223794 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.223805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.223818 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.223829 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.325995 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.326041 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.326050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.326066 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.326078 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.429096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.429168 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.429182 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.429201 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.429213 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.532252 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.532294 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.532302 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.532316 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.532325 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.634464 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.634494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.634745 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.634765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.634774 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.737664 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.737719 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.737730 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.737749 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.737764 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.776758 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.776803 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:50 crc kubenswrapper[4706]: E1127 07:09:50.776912 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.776933 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:50 crc kubenswrapper[4706]: E1127 07:09:50.777048 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:50 crc kubenswrapper[4706]: E1127 07:09:50.777286 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.840895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.840930 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.840939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.840951 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.840960 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.944066 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.944109 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.944120 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.944135 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:50 crc kubenswrapper[4706]: I1127 07:09:50.944146 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:50Z","lastTransitionTime":"2025-11-27T07:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.047440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.047482 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.047493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.047509 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.047522 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.150628 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.150690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.150709 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.150734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.150753 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.253943 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.254038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.254058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.254112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.254130 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.357794 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.357850 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.357859 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.357877 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.357888 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.461181 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.461263 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.461281 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.461306 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.461323 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.564461 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.564496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.564507 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.564540 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.564549 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.667305 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.667340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.667352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.667368 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.667377 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.770430 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.770473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.770481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.770495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.770504 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.776024 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:51 crc kubenswrapper[4706]: E1127 07:09:51.776125 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.777275 4706 scope.go:117] "RemoveContainer" containerID="9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.873805 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.873842 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.873854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.873869 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.873882 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.975661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.975706 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.975724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.975807 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:51 crc kubenswrapper[4706]: I1127 07:09:51.975824 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:51Z","lastTransitionTime":"2025-11-27T07:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.078372 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.078424 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.078435 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.078453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.078465 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.170251 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/2.log" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.173326 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.173876 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.180518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.180557 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.180567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.180582 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.180592 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.197562 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.214395 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.227194 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.244400 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.256387 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.267746 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.280000 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.282860 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.282887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.282895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.282909 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.282920 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.289604 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.301288 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.314603 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.324971 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.334870 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.347697 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:45Z\\\",\\\"message\\\":\\\"2025-11-27T07:08:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5\\\\n2025-11-27T07:08:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5 to /host/opt/cni/bin/\\\\n2025-11-27T07:09:00Z [verbose] multus-daemon started\\\\n2025-11-27T07:09:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T07:09:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.358281 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.376203 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.384705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.384750 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.384761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.384775 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.384785 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.389573 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.400764 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.486795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.486855 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.486868 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.486886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.486914 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.591201 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.591257 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.591266 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.591279 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.591290 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.693450 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.693496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.693504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.693519 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.693530 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.776771 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.776817 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.776870 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:52 crc kubenswrapper[4706]: E1127 07:09:52.776914 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:52 crc kubenswrapper[4706]: E1127 07:09:52.777071 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:52 crc kubenswrapper[4706]: E1127 07:09:52.777105 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.795955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.795977 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.795986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.795997 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.796006 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.898420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.898724 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.898815 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.898894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:52 crc kubenswrapper[4706]: I1127 07:09:52.898962 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:52Z","lastTransitionTime":"2025-11-27T07:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.001215 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.001289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.001297 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.001311 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.001320 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.103698 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.103736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.103744 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.103757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.103765 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.177501 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/3.log" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.178185 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/2.log" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.180796 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" exitCode=1 Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.180843 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.180887 4706 scope.go:117] "RemoveContainer" containerID="9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.181877 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:09:53 crc kubenswrapper[4706]: E1127 07:09:53.182144 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.199839 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.206259 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.206307 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.206317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.206336 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.206347 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.212533 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.222497 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.236709 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.247329 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.258486 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:45Z\\\",\\\"message\\\":\\\"2025-11-27T07:08:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5\\\\n2025-11-27T07:08:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5 to /host/opt/cni/bin/\\\\n2025-11-27T07:09:00Z [verbose] multus-daemon started\\\\n2025-11-27T07:09:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T07:09:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.274202 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.287654 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.299679 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.308723 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.308766 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.308777 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.308795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.308806 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.311272 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.323739 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.338550 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.350132 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.362087 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.378063 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9318d55954b1e8ce68d9ee2ffa29e862f8105f88c24db51858418f2ff7496628\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1127 07:09:24.649907 6363 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 07:09:24.649926 6363 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:24.649906 6363 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1127 07:09:24.649927 6363 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:24.649944 6363 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:24.649973 6363 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:24.649980 6363 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:24.650017 6363 factory.go:656] Stopping watch factory\\\\nI1127 07:09:24.650038 6363 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:24.650043 6363 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:24.650047 6363 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:24.650247 6363 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:24.650320 6363 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:24.650354 6363 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:24.650377 6363 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:24.650450 6363 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:52Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI1127 07:09:52.615431 6725 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:52.615460 6725 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:52.615468 6725 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:52.615481 6725 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:52.615500 6725 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:52.615521 6725 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:52.615521 6725 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:52.615528 6725 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:52.615629 6725 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:52.615662 6725 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1127 07:09:52.615678 6725 factory.go:656] Stopping watch factory\\\\nI1127 07:09:52.615691 6725 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 07:09:52.615699 6725 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:52.615730 6725 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:52.615747 6725 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:52.615804 6725 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.388850 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.399089 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.411026 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.411235 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.411373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.411639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.411836 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.514142 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.514373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.514459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.514530 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.514594 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.617155 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.617441 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.617755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.618025 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.618178 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.721266 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.721317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.721333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.721351 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.721365 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.775840 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:53 crc kubenswrapper[4706]: E1127 07:09:53.776535 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.823800 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.823827 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.823837 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.823853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.823863 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.926165 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.926189 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.926197 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.926209 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:53 crc kubenswrapper[4706]: I1127 07:09:53.926219 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:53Z","lastTransitionTime":"2025-11-27T07:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.027986 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.028097 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.028116 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.028140 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.028157 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.130323 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.130355 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.130363 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.130377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.130386 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.186215 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/3.log" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.189917 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.190067 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.201544 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.213625 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.233304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.233335 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.233343 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.233357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.233366 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.233696 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.245726 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.257626 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.276052 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:52Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI1127 07:09:52.615431 6725 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:52.615460 6725 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:52.615468 6725 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:52.615481 6725 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:52.615500 6725 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:52.615521 6725 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:52.615521 6725 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:52.615528 6725 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:52.615629 6725 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:52.615662 6725 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1127 07:09:52.615678 6725 factory.go:656] Stopping watch factory\\\\nI1127 07:09:52.615691 6725 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 07:09:52.615699 6725 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:52.615730 6725 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:52.615747 6725 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:52.615804 6725 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.288492 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.303487 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.321802 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.335746 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.335776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.335785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.335798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.335807 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.337621 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.352408 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.371296 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:45Z\\\",\\\"message\\\":\\\"2025-11-27T07:08:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5\\\\n2025-11-27T07:08:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5 to /host/opt/cni/bin/\\\\n2025-11-27T07:09:00Z [verbose] multus-daemon started\\\\n2025-11-27T07:09:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T07:09:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.386117 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.405646 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.423460 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.437886 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.437947 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.437959 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.437976 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.437985 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.441504 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.450324 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.540217 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.540274 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.540285 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.540301 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.540313 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.643663 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.643717 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.643734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.643755 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.643771 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.746402 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.746483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.746507 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.746538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.746560 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.776435 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.776488 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.776491 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.776568 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.776699 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.776817 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.848711 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.848757 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.848772 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.848790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.848802 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.877455 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.877513 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.877524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.877537 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.877546 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.888915 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.892012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.892047 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.892056 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.892089 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.892097 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.902969 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.906279 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.906314 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.906323 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.906337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.906347 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.919027 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.922633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.922672 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.922685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.922701 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.922714 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.932931 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.936236 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.936289 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.936299 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.936315 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.936325 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.948753 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ab140b-da2b-4616-bf57-a58441a0c7b5\\\",\\\"systemUUID\\\":\\\"70cebc65-54fc-48ee-bf03-1e5c2a992650\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:54 crc kubenswrapper[4706]: E1127 07:09:54.948880 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.951369 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.951466 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.951529 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.951601 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:54 crc kubenswrapper[4706]: I1127 07:09:54.951672 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:54Z","lastTransitionTime":"2025-11-27T07:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.054708 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.054754 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.054765 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.054783 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.054795 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.157091 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.157129 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.157137 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.157149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.157158 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.258984 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.259017 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.259025 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.259038 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.259049 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.361801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.361864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.361882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.361908 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.361925 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.464149 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.464213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.464263 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.464286 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.464300 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.567148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.567192 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.567203 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.567234 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.567247 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.669310 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.669350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.669358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.669373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.669383 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.771357 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.771408 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.771423 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.771443 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.771457 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.776634 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:55 crc kubenswrapper[4706]: E1127 07:09:55.776739 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.873454 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.873510 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.873524 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.873538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.873549 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.976324 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.976396 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.976420 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.976440 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:55 crc kubenswrapper[4706]: I1127 07:09:55.976455 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:55Z","lastTransitionTime":"2025-11-27T07:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.079134 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.079649 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.079661 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.079677 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.079688 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.182292 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.182352 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.182364 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.182382 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.182394 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.284824 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.284858 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.284867 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.284880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.284888 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.386780 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.386821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.386830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.386844 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.386856 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.488567 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.488611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.488620 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.488636 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.488648 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.591624 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.591671 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.591679 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.591693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.591703 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.693882 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.693983 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.693996 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.694012 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.694023 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.776315 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.776332 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:56 crc kubenswrapper[4706]: E1127 07:09:56.776600 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.776346 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:56 crc kubenswrapper[4706]: E1127 07:09:56.776709 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:56 crc kubenswrapper[4706]: E1127 07:09:56.776764 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.793820 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2df88e1-8d99-4164-9919-1f4f2890fa26\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee12edf88dd0055569fb720346b6e74d5938a4eec3af0d67b30aa469d1a3117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50564197c7ca30255d171bc5f4ef831d0aaeb1f3428f276ae3afda27349ef578\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5559137a7133ac65f0d83b0c802d344a598db37a303c37deb1d781ebb0e6ce8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.796512 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.796563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.796576 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.796596 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.796610 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.811528 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"324368bd-b292-44dc-967f-e73512448d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f464e3bf35c6613a41ca28839fa78ab8a378f570496993fa9fd0a096a48bbc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba62aa82a5c2df5a9836a31a96409db4be55f7b5ff80569c9d8c315068cd9898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b1df2b82de4b13ba60f5603c5371a01744253addde69028ec2cf20f201fbe86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44a5d0adf7d4749fb426c53f287f1a622a20695ebe1cfa961af4babc007890c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.829285 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxrlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zxp2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.843941 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89d12ee-9aae-48da-a7a6-130c7186f506\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3c8be62455ea514b40d2f88ee0a4fc1908e33d3e5a40b524eccadf5ade48251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv675\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c44hv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.859562 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m645h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e33550a-5308-4e22-9c81-8a12e97f44ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:45Z\\\",\\\"message\\\":\\\"2025-11-27T07:08:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5\\\\n2025-11-27T07:08:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bf59041d-5a3f-47cd-b932-bf22e2c757c5 to /host/opt/cni/bin/\\\\n2025-11-27T07:09:00Z [verbose] multus-daemon started\\\\n2025-11-27T07:09:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T07:09:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ns2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m645h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.872368 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cb8d10b-fddc-42e4-9181-6366603ea340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5edd8e9245d609e5702b2494b973d5a3d0a4f721d7c85acfe2abdd7135fa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98a24f23dedb0075c90e5fe8d99dfd961edd0f25534cadb540258b4845261da2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b29da0499cf15c0fea238d825d041dcb6a0d56876bad7c139e6df0c5c92cd9bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24f54bcce0c384e993e4cdd62cb9f65a6de84c25171a7278fc1a84eabc8d33c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b21ba3a5a17c8dc6a5e46d73cb133c1eab7e4f6b197dd2b88d359feb77c5f767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b4ce68f1933a3acdc0fb8240e93d861d6d4df3e92cf4cb8b78b5aaea86cc4dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a59c49ae2469ed03033c1a959cb798c22292c8d067346c080ea8325c78d1bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jlfzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p9dt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.885071 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ee10a2937909a216f2dd8bdd06508a3bff1746d8733fef36fac627e811a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.898569 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.898597 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.898605 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.898618 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.898627 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:56Z","lastTransitionTime":"2025-11-27T07:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.902886 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.914143 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://043b20cee282a8700df6384c13eab88da698376d0f9a440454f7b6111511d5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb914e893d8d828efe651dfba804db43521fe820366b908d4a81afdc8f88f04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.923821 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mrfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4272330e-1068-42ec-92d1-3a3503adaef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790b673838af5a3829ceb2ad4f08b3d89b157ab77f3eba7d4e65064661badc13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8hdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mrfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.935144 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c7d86c1bae008152676ef3d294a1db3186286b39afb1f6bb41464513788c010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.949156 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.959669 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.975807 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1b93703-3a71-49b6-bff6-b4d314006ddd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T07:09:52Z\\\",\\\"message\\\":\\\"Node event handler 7 for removal\\\\nI1127 07:09:52.615431 6725 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 07:09:52.615460 6725 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1127 07:09:52.615468 6725 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1127 07:09:52.615481 6725 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 07:09:52.615500 6725 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 07:09:52.615521 6725 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 07:09:52.615521 6725 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 07:09:52.615528 6725 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 07:09:52.615629 6725 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1127 07:09:52.615662 6725 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1127 07:09:52.615678 6725 factory.go:656] Stopping watch factory\\\\nI1127 07:09:52.615691 6725 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 07:09:52.615699 6725 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1127 07:09:52.615730 6725 ovnkube.go:599] Stopped ovnkube\\\\nI1127 07:09:52.615747 6725 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1127 07:09:52.615804 6725 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:09:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s94lx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:56 crc kubenswrapper[4706]: I1127 07:09:56.995105 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2hssd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026dfb3b-77c4-48bf-9402-a6fedc63f55a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d495a79d5f71eb586b3702971720410ff7957929a879c9807d0052f00116d26d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n5t6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2hssd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.001166 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.001213 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.001244 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.001262 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.001272 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.005600 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6071ddb-03d9-4187-9e75-1abddee42d27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b11aa1c44c391b3187dc86c05ef2f2927730eed1e12629563d6394bda8064315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fd23b562d0b80314f4baa4f26c6fa7e4f71feedb97a1116a501013123900732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2mpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nf57s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.017654 4706 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee814c58-bb88-4206-bf64-22abdbdeae78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T07:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T07:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1127 07:08:50.561714 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 07:08:50.565347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3806981535/tls.crt::/tmp/serving-cert-3806981535/tls.key\\\\\\\"\\\\nI1127 07:08:56.739835 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 07:08:56.743571 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 07:08:56.743594 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 07:08:56.743617 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 07:08:56.743624 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 07:08:56.749599 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1127 07:08:56.749620 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1127 07:08:56.749629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749635 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 07:08:56.749639 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 07:08:56.749644 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 07:08:56.749649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 07:08:56.749656 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1127 07:08:56.754775 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T07:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T07:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T07:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T07:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T07:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.103792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.104079 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.104164 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.104262 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.104340 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.207377 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.207483 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.207495 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.207880 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.207899 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.311906 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.311982 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.312020 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.312052 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.312082 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.414743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.414810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.414823 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.414846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.414864 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.517004 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.517051 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.517060 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.517080 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.517093 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.620761 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.620835 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.620855 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.620884 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.620906 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.724409 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.724456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.724465 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.724491 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.724503 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.776408 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:57 crc kubenswrapper[4706]: E1127 07:09:57.776599 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.827411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.827475 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.827492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.827520 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.827536 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.931094 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.931139 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.931148 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.931166 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:57 crc kubenswrapper[4706]: I1127 07:09:57.931175 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:57Z","lastTransitionTime":"2025-11-27T07:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.034686 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.034753 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.034792 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.034819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.034837 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.137611 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.137658 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.137668 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.137685 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.137699 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.239694 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.239751 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.239763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.239778 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.239788 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.342939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.342993 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.343002 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.343019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.343030 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.446319 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.446389 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.446411 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.446443 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.446470 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.550732 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.550819 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.550846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.550887 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.550917 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.654688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.654762 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.654776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.654801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.654851 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.758146 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.758288 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.758323 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.758359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.758385 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.776211 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:09:58 crc kubenswrapper[4706]: E1127 07:09:58.776433 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.776565 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.776590 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:09:58 crc kubenswrapper[4706]: E1127 07:09:58.776843 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:09:58 crc kubenswrapper[4706]: E1127 07:09:58.776922 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.862298 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.862361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.862374 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.862394 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.862407 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.965518 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.965551 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.965561 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.965578 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:58 crc kubenswrapper[4706]: I1127 07:09:58.965588 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:58Z","lastTransitionTime":"2025-11-27T07:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.067871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.067913 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.067921 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.067937 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.067946 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.170286 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.170338 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.170354 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.170373 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.170385 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.272451 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.272504 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.272514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.272532 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.272545 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.375256 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.375324 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.375337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.375359 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.375373 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.478011 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.478069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.478081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.478104 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.478114 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.581301 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.581340 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.581348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.581361 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.581370 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.689570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.689894 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.689906 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.689922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.689932 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.776147 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:09:59 crc kubenswrapper[4706]: E1127 07:09:59.776962 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.792614 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.792917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.793350 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.793734 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.794092 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.897419 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.897496 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.897516 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.897549 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:09:59 crc kubenswrapper[4706]: I1127 07:09:59.897570 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:09:59Z","lastTransitionTime":"2025-11-27T07:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.000286 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.000349 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.000367 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.000400 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.000423 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.102802 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.102890 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.102910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.102939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.102959 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.206432 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.206521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.206534 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.206577 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.206595 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.309555 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.309595 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.309604 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.309619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.309629 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.412688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.412733 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.412743 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.412763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.412775 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.516216 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.516314 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.516333 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.516348 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.516357 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.591516 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.591588 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.591625 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.591669 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.59164641 +0000 UTC m=+148.481237220 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.591701 4706 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.591736 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.591728012 +0000 UTC m=+148.481318822 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.591893 4706 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.592072 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.592039742 +0000 UTC m=+148.481630582 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.619864 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.619922 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.619935 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.619958 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.619973 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.693149 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.693194 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693351 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693370 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693382 4706 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693392 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693430 4706 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693452 4706 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693435 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.693418338 +0000 UTC m=+148.583009148 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.693526 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.693507361 +0000 UTC m=+148.583098201 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.722615 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.722654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.722662 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.722676 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.722685 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.776179 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.776247 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.776326 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.776178 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.776514 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:00 crc kubenswrapper[4706]: E1127 07:10:00.776634 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.825061 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.825101 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.825110 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.825125 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.825135 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.928057 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.928112 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.928127 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.928151 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:00 crc kubenswrapper[4706]: I1127 07:10:00.928169 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:00Z","lastTransitionTime":"2025-11-27T07:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.031027 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.031087 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.031102 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.031121 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.031133 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.133942 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.133987 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.133998 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.134031 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.134041 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.236840 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.236876 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.236885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.236898 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.236907 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.338874 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.338910 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.338924 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.338938 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.338947 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.441731 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.441767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.441776 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.441790 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.441799 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.544247 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.544482 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.544759 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.545003 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.545166 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.647599 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.647640 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.647653 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.647693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.647702 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.751014 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.751050 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.751058 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.751072 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.751081 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.776679 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:01 crc kubenswrapper[4706]: E1127 07:10:01.776830 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.853446 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.853693 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.853785 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.853901 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.857437 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.962885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.962917 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.962925 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.962939 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:01 crc kubenswrapper[4706]: I1127 07:10:01.962948 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:01Z","lastTransitionTime":"2025-11-27T07:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.065489 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.065527 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.065538 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.065594 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.065608 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.167794 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.167829 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.167838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.167853 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.167863 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.270688 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.270720 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.270728 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.270740 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.270749 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.373563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.373617 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.373634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.373656 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.373673 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.476198 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.476270 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.476284 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.476304 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.476318 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.579725 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.579801 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.579821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.579846 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.579862 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.682639 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.682680 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.682690 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.682705 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.682715 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.776553 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.776728 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.776845 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:02 crc kubenswrapper[4706]: E1127 07:10:02.776953 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:02 crc kubenswrapper[4706]: E1127 07:10:02.777060 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:02 crc kubenswrapper[4706]: E1127 07:10:02.777289 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.785069 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.785096 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.785108 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.785126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.785143 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.788555 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.887456 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.887492 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.887501 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.887514 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.887523 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.989764 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.989803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.989814 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.989830 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:02 crc kubenswrapper[4706]: I1127 07:10:02.989841 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:02Z","lastTransitionTime":"2025-11-27T07:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.092667 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.092727 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.092736 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.092752 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.092760 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.195250 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.195294 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.195317 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.195337 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.195352 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.297479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.297511 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.297521 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.297536 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.297547 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.400019 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.400070 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.400084 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.400103 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.400117 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.502767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.502810 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.502821 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.502838 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.502850 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.605048 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.605126 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.605138 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.605183 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.605196 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.708397 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.708459 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.708473 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.708494 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.708509 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.776351 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:03 crc kubenswrapper[4706]: E1127 07:10:03.776596 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.811763 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.811795 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.811803 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.811817 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.811826 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.913747 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.913793 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.913804 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.913823 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:03 crc kubenswrapper[4706]: I1127 07:10:03.913835 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:03Z","lastTransitionTime":"2025-11-27T07:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.015788 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.016081 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.016216 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.016407 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.016540 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.118388 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.118463 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.118477 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.118493 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.118503 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.221453 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.221546 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.221563 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.221584 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.221600 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.324368 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.324439 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.324479 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.324511 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.324534 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.427570 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.427610 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.427619 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.427633 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.427644 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.529885 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.529926 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.529940 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.529955 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.529965 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.632798 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.632903 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.632927 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.632961 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.632983 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.736199 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.736313 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.736332 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.736358 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.736375 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.776190 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.776410 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:04 crc kubenswrapper[4706]: E1127 07:10:04.776621 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.776708 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:04 crc kubenswrapper[4706]: E1127 07:10:04.777103 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:04 crc kubenswrapper[4706]: E1127 07:10:04.777456 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.838767 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.838820 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.838831 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.838849 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.838861 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.941167 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.941238 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.941251 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.941272 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:04 crc kubenswrapper[4706]: I1127 07:10:04.941283 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:04Z","lastTransitionTime":"2025-11-27T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.043598 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.043634 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.043642 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.043654 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.043662 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:05Z","lastTransitionTime":"2025-11-27T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.146791 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.146854 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.146871 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.146895 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.146914 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:05Z","lastTransitionTime":"2025-11-27T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.161481 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.161526 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.161543 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.161566 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.161584 4706 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T07:10:05Z","lastTransitionTime":"2025-11-27T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.215000 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd"] Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.215394 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.216851 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.217402 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.217598 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.218650 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.230579 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=3.230562389 podStartE2EDuration="3.230562389s" podCreationTimestamp="2025-11-27 07:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.230543809 +0000 UTC m=+89.120134649" watchObservedRunningTime="2025-11-27 07:10:05.230562389 +0000 UTC m=+89.120153199" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.238699 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.238738 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.238781 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.238807 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.238832 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.257327 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.257307051 podStartE2EDuration="1m7.257307051s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.246384165 +0000 UTC m=+89.135975015" watchObservedRunningTime="2025-11-27 07:10:05.257307051 +0000 UTC m=+89.146897861" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.271467 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.271448035 podStartE2EDuration="41.271448035s" podCreationTimestamp="2025-11-27 07:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.258326432 +0000 UTC m=+89.147917282" watchObservedRunningTime="2025-11-27 07:10:05.271448035 +0000 UTC m=+89.161038855" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.302596 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podStartSLOduration=68.302577402 podStartE2EDuration="1m8.302577402s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.290931544 +0000 UTC m=+89.180522354" watchObservedRunningTime="2025-11-27 07:10:05.302577402 +0000 UTC m=+89.192168212" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.319149 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-m645h" podStartSLOduration=68.319129351 podStartE2EDuration="1m8.319129351s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.302796689 +0000 UTC m=+89.192387499" watchObservedRunningTime="2025-11-27 07:10:05.319129351 +0000 UTC m=+89.208720161" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.328733 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-p9dt7" podStartSLOduration=68.328716136 podStartE2EDuration="1m8.328716136s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.319047438 +0000 UTC m=+89.208638248" watchObservedRunningTime="2025-11-27 07:10:05.328716136 +0000 UTC m=+89.218306946" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.339286 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.339333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.339362 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.339388 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.339407 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.339453 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.339463 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.340423 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.344758 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.356493 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/efb658b5-0ffe-4fb4-b087-88a7822ab8b9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qhnzd\" (UID: \"efb658b5-0ffe-4fb4-b087-88a7822ab8b9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.388294 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5mrfn" podStartSLOduration=68.388273286 podStartE2EDuration="1m8.388273286s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.388100551 +0000 UTC m=+89.277691371" watchObservedRunningTime="2025-11-27 07:10:05.388273286 +0000 UTC m=+89.277864096" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.405146 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.405124324 podStartE2EDuration="1m9.405124324s" podCreationTimestamp="2025-11-27 07:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.404484894 +0000 UTC m=+89.294075724" watchObservedRunningTime="2025-11-27 07:10:05.405124324 +0000 UTC m=+89.294715144" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.458063 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2hssd" podStartSLOduration=68.45804549 podStartE2EDuration="1m8.45804549s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.457821383 +0000 UTC m=+89.347412193" watchObservedRunningTime="2025-11-27 07:10:05.45804549 +0000 UTC m=+89.347636300" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.543301 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" Nov 27 07:10:05 crc kubenswrapper[4706]: I1127 07:10:05.776710 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:05 crc kubenswrapper[4706]: E1127 07:10:05.776827 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:06 crc kubenswrapper[4706]: I1127 07:10:06.230577 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" event={"ID":"efb658b5-0ffe-4fb4-b087-88a7822ab8b9","Type":"ContainerStarted","Data":"4fa8fd27cd8795e081b09af1937fce6fc193f6d87acf1b9921d73d12d863586b"} Nov 27 07:10:06 crc kubenswrapper[4706]: I1127 07:10:06.230633 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" event={"ID":"efb658b5-0ffe-4fb4-b087-88a7822ab8b9","Type":"ContainerStarted","Data":"34a9630452de7b3d857bcaa3731fd95711ec04368b21f4c217520b3842fb4fbc"} Nov 27 07:10:06 crc kubenswrapper[4706]: I1127 07:10:06.243382 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nf57s" podStartSLOduration=68.243367044 podStartE2EDuration="1m8.243367044s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:05.468862773 +0000 UTC m=+89.358453603" watchObservedRunningTime="2025-11-27 07:10:06.243367044 +0000 UTC m=+90.132957854" Nov 27 07:10:06 crc kubenswrapper[4706]: I1127 07:10:06.776313 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:06 crc kubenswrapper[4706]: I1127 07:10:06.776646 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:06 crc kubenswrapper[4706]: I1127 07:10:06.776714 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:06 crc kubenswrapper[4706]: E1127 07:10:06.778983 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:06 crc kubenswrapper[4706]: E1127 07:10:06.779113 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:06 crc kubenswrapper[4706]: E1127 07:10:06.779246 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:07 crc kubenswrapper[4706]: I1127 07:10:07.776412 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:07 crc kubenswrapper[4706]: E1127 07:10:07.776657 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:08 crc kubenswrapper[4706]: I1127 07:10:08.776885 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:08 crc kubenswrapper[4706]: I1127 07:10:08.777037 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:08 crc kubenswrapper[4706]: I1127 07:10:08.777090 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:08 crc kubenswrapper[4706]: E1127 07:10:08.777350 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:08 crc kubenswrapper[4706]: E1127 07:10:08.777631 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:08 crc kubenswrapper[4706]: E1127 07:10:08.777688 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:09 crc kubenswrapper[4706]: I1127 07:10:09.775746 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:09 crc kubenswrapper[4706]: E1127 07:10:09.776230 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:09 crc kubenswrapper[4706]: I1127 07:10:09.776597 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:10:09 crc kubenswrapper[4706]: E1127 07:10:09.776786 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:10:10 crc kubenswrapper[4706]: I1127 07:10:10.776949 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:10 crc kubenswrapper[4706]: I1127 07:10:10.777067 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:10 crc kubenswrapper[4706]: E1127 07:10:10.777283 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:10 crc kubenswrapper[4706]: I1127 07:10:10.777412 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:10 crc kubenswrapper[4706]: E1127 07:10:10.777556 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:10 crc kubenswrapper[4706]: E1127 07:10:10.777661 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:11 crc kubenswrapper[4706]: I1127 07:10:11.776084 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:11 crc kubenswrapper[4706]: E1127 07:10:11.776989 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:12 crc kubenswrapper[4706]: I1127 07:10:12.776352 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:12 crc kubenswrapper[4706]: I1127 07:10:12.776436 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:12 crc kubenswrapper[4706]: E1127 07:10:12.776572 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:12 crc kubenswrapper[4706]: E1127 07:10:12.776760 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:12 crc kubenswrapper[4706]: I1127 07:10:12.777104 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:12 crc kubenswrapper[4706]: E1127 07:10:12.777265 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:13 crc kubenswrapper[4706]: I1127 07:10:13.776110 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:13 crc kubenswrapper[4706]: E1127 07:10:13.776280 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:14 crc kubenswrapper[4706]: I1127 07:10:14.776952 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:14 crc kubenswrapper[4706]: E1127 07:10:14.777970 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:14 crc kubenswrapper[4706]: I1127 07:10:14.777020 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:14 crc kubenswrapper[4706]: E1127 07:10:14.778351 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:14 crc kubenswrapper[4706]: I1127 07:10:14.776952 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:14 crc kubenswrapper[4706]: E1127 07:10:14.778662 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:15 crc kubenswrapper[4706]: I1127 07:10:15.776245 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:15 crc kubenswrapper[4706]: E1127 07:10:15.776529 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:15 crc kubenswrapper[4706]: I1127 07:10:15.872856 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:15 crc kubenswrapper[4706]: E1127 07:10:15.873036 4706 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:10:15 crc kubenswrapper[4706]: E1127 07:10:15.873117 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs podName:ab5bf95b-5c0d-4804-bc97-43b79f6f2fea nodeName:}" failed. No retries permitted until 2025-11-27 07:11:19.873093149 +0000 UTC m=+163.762683969 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs") pod "network-metrics-daemon-zxp2h" (UID: "ab5bf95b-5c0d-4804-bc97-43b79f6f2fea") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 07:10:16 crc kubenswrapper[4706]: I1127 07:10:16.776377 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:16 crc kubenswrapper[4706]: I1127 07:10:16.776571 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:16 crc kubenswrapper[4706]: E1127 07:10:16.777242 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:16 crc kubenswrapper[4706]: I1127 07:10:16.777280 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:16 crc kubenswrapper[4706]: E1127 07:10:16.777640 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:16 crc kubenswrapper[4706]: E1127 07:10:16.777738 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:17 crc kubenswrapper[4706]: I1127 07:10:17.776820 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:17 crc kubenswrapper[4706]: E1127 07:10:17.777051 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:18 crc kubenswrapper[4706]: I1127 07:10:18.776023 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:18 crc kubenswrapper[4706]: E1127 07:10:18.776297 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:18 crc kubenswrapper[4706]: I1127 07:10:18.776062 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:18 crc kubenswrapper[4706]: I1127 07:10:18.777268 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:18 crc kubenswrapper[4706]: E1127 07:10:18.777484 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:18 crc kubenswrapper[4706]: E1127 07:10:18.777658 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:19 crc kubenswrapper[4706]: I1127 07:10:19.776020 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:19 crc kubenswrapper[4706]: E1127 07:10:19.776191 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:20 crc kubenswrapper[4706]: I1127 07:10:20.776152 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:20 crc kubenswrapper[4706]: I1127 07:10:20.776266 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:20 crc kubenswrapper[4706]: I1127 07:10:20.776152 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:20 crc kubenswrapper[4706]: E1127 07:10:20.776386 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:20 crc kubenswrapper[4706]: E1127 07:10:20.776501 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:20 crc kubenswrapper[4706]: E1127 07:10:20.776650 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:21 crc kubenswrapper[4706]: I1127 07:10:21.775825 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:21 crc kubenswrapper[4706]: E1127 07:10:21.776330 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:21 crc kubenswrapper[4706]: I1127 07:10:21.776599 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:10:21 crc kubenswrapper[4706]: E1127 07:10:21.776749 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhhb_openshift-ovn-kubernetes(d1b93703-3a71-49b6-bff6-b4d314006ddd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" Nov 27 07:10:22 crc kubenswrapper[4706]: I1127 07:10:22.775824 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:22 crc kubenswrapper[4706]: I1127 07:10:22.775949 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:22 crc kubenswrapper[4706]: I1127 07:10:22.776011 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:22 crc kubenswrapper[4706]: E1127 07:10:22.776119 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:22 crc kubenswrapper[4706]: E1127 07:10:22.776313 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:22 crc kubenswrapper[4706]: E1127 07:10:22.776480 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:23 crc kubenswrapper[4706]: I1127 07:10:23.775947 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:23 crc kubenswrapper[4706]: E1127 07:10:23.776211 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:23 crc kubenswrapper[4706]: I1127 07:10:23.793520 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qhnzd" podStartSLOduration=86.793501316 podStartE2EDuration="1m26.793501316s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:06.243685104 +0000 UTC m=+90.133275914" watchObservedRunningTime="2025-11-27 07:10:23.793501316 +0000 UTC m=+107.683092126" Nov 27 07:10:23 crc kubenswrapper[4706]: I1127 07:10:23.794300 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 27 07:10:24 crc kubenswrapper[4706]: I1127 07:10:24.775938 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:24 crc kubenswrapper[4706]: E1127 07:10:24.776096 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:24 crc kubenswrapper[4706]: I1127 07:10:24.776300 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:24 crc kubenswrapper[4706]: E1127 07:10:24.776463 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:24 crc kubenswrapper[4706]: I1127 07:10:24.776575 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:24 crc kubenswrapper[4706]: E1127 07:10:24.776708 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:25 crc kubenswrapper[4706]: I1127 07:10:25.776347 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:25 crc kubenswrapper[4706]: E1127 07:10:25.776520 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:26 crc kubenswrapper[4706]: I1127 07:10:26.776029 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:26 crc kubenswrapper[4706]: I1127 07:10:26.776068 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:26 crc kubenswrapper[4706]: I1127 07:10:26.777398 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:26 crc kubenswrapper[4706]: E1127 07:10:26.777494 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:26 crc kubenswrapper[4706]: E1127 07:10:26.777589 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:26 crc kubenswrapper[4706]: E1127 07:10:26.777348 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:26 crc kubenswrapper[4706]: I1127 07:10:26.817682 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.817664354 podStartE2EDuration="3.817664354s" podCreationTimestamp="2025-11-27 07:10:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:26.816551109 +0000 UTC m=+110.706141939" watchObservedRunningTime="2025-11-27 07:10:26.817664354 +0000 UTC m=+110.707255164" Nov 27 07:10:27 crc kubenswrapper[4706]: I1127 07:10:27.775965 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:27 crc kubenswrapper[4706]: E1127 07:10:27.776084 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:28 crc kubenswrapper[4706]: I1127 07:10:28.776796 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:28 crc kubenswrapper[4706]: I1127 07:10:28.776847 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:28 crc kubenswrapper[4706]: I1127 07:10:28.776989 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:28 crc kubenswrapper[4706]: E1127 07:10:28.777145 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:28 crc kubenswrapper[4706]: E1127 07:10:28.777340 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:28 crc kubenswrapper[4706]: E1127 07:10:28.777489 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:29 crc kubenswrapper[4706]: I1127 07:10:29.775814 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:29 crc kubenswrapper[4706]: E1127 07:10:29.775944 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:30 crc kubenswrapper[4706]: I1127 07:10:30.776511 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:30 crc kubenswrapper[4706]: I1127 07:10:30.776547 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:30 crc kubenswrapper[4706]: I1127 07:10:30.776567 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:30 crc kubenswrapper[4706]: E1127 07:10:30.776660 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:30 crc kubenswrapper[4706]: E1127 07:10:30.776739 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:30 crc kubenswrapper[4706]: E1127 07:10:30.776834 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:31 crc kubenswrapper[4706]: I1127 07:10:31.776330 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:31 crc kubenswrapper[4706]: E1127 07:10:31.776504 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.317568 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/1.log" Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.318134 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/0.log" Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.318215 4706 generic.go:334] "Generic (PLEG): container finished" podID="0e33550a-5308-4e22-9c81-8a12e97f44ef" containerID="5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6" exitCode=1 Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.318304 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerDied","Data":"5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6"} Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.318361 4706 scope.go:117] "RemoveContainer" containerID="f90ec0b38b4c3da1b5ffb123919153a8593c3bcd9f21a9b5d39bf0feabc43df7" Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.318978 4706 scope.go:117] "RemoveContainer" containerID="5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6" Nov 27 07:10:32 crc kubenswrapper[4706]: E1127 07:10:32.319262 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-m645h_openshift-multus(0e33550a-5308-4e22-9c81-8a12e97f44ef)\"" pod="openshift-multus/multus-m645h" podUID="0e33550a-5308-4e22-9c81-8a12e97f44ef" Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.776733 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.776798 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:32 crc kubenswrapper[4706]: I1127 07:10:32.776738 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:32 crc kubenswrapper[4706]: E1127 07:10:32.776906 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:32 crc kubenswrapper[4706]: E1127 07:10:32.777014 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:32 crc kubenswrapper[4706]: E1127 07:10:32.777091 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:33 crc kubenswrapper[4706]: I1127 07:10:33.323204 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/1.log" Nov 27 07:10:33 crc kubenswrapper[4706]: I1127 07:10:33.776005 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:33 crc kubenswrapper[4706]: E1127 07:10:33.776820 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:33 crc kubenswrapper[4706]: I1127 07:10:33.777520 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.328758 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/3.log" Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.331679 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerStarted","Data":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.332132 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.556557 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podStartSLOduration=97.556537242 podStartE2EDuration="1m37.556537242s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:10:34.363396026 +0000 UTC m=+118.252986846" watchObservedRunningTime="2025-11-27 07:10:34.556537242 +0000 UTC m=+118.446128052" Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.557366 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zxp2h"] Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.557457 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:34 crc kubenswrapper[4706]: E1127 07:10:34.557552 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.775962 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.775977 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:34 crc kubenswrapper[4706]: E1127 07:10:34.776072 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:34 crc kubenswrapper[4706]: I1127 07:10:34.776085 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:34 crc kubenswrapper[4706]: E1127 07:10:34.776130 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:34 crc kubenswrapper[4706]: E1127 07:10:34.776298 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:35 crc kubenswrapper[4706]: I1127 07:10:35.776322 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:35 crc kubenswrapper[4706]: E1127 07:10:35.776474 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:36 crc kubenswrapper[4706]: I1127 07:10:36.776576 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:36 crc kubenswrapper[4706]: I1127 07:10:36.776597 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:36 crc kubenswrapper[4706]: E1127 07:10:36.776637 4706 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 27 07:10:36 crc kubenswrapper[4706]: I1127 07:10:36.776650 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:36 crc kubenswrapper[4706]: E1127 07:10:36.777625 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:36 crc kubenswrapper[4706]: E1127 07:10:36.777678 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:36 crc kubenswrapper[4706]: E1127 07:10:36.777721 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:36 crc kubenswrapper[4706]: E1127 07:10:36.892719 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 07:10:37 crc kubenswrapper[4706]: I1127 07:10:37.776629 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:37 crc kubenswrapper[4706]: E1127 07:10:37.777203 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:38 crc kubenswrapper[4706]: I1127 07:10:38.776458 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:38 crc kubenswrapper[4706]: I1127 07:10:38.776535 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:38 crc kubenswrapper[4706]: E1127 07:10:38.776591 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:38 crc kubenswrapper[4706]: I1127 07:10:38.776629 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:38 crc kubenswrapper[4706]: E1127 07:10:38.776785 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:38 crc kubenswrapper[4706]: E1127 07:10:38.776888 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:39 crc kubenswrapper[4706]: I1127 07:10:39.776855 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:39 crc kubenswrapper[4706]: E1127 07:10:39.777050 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:40 crc kubenswrapper[4706]: I1127 07:10:40.775983 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:40 crc kubenswrapper[4706]: I1127 07:10:40.776029 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:40 crc kubenswrapper[4706]: I1127 07:10:40.776126 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:40 crc kubenswrapper[4706]: E1127 07:10:40.776259 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:40 crc kubenswrapper[4706]: E1127 07:10:40.776326 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:40 crc kubenswrapper[4706]: E1127 07:10:40.776435 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:41 crc kubenswrapper[4706]: I1127 07:10:41.775882 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:41 crc kubenswrapper[4706]: E1127 07:10:41.776077 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:41 crc kubenswrapper[4706]: E1127 07:10:41.894257 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 07:10:42 crc kubenswrapper[4706]: I1127 07:10:42.776547 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:42 crc kubenswrapper[4706]: I1127 07:10:42.776603 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:42 crc kubenswrapper[4706]: I1127 07:10:42.776737 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:42 crc kubenswrapper[4706]: E1127 07:10:42.776795 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:42 crc kubenswrapper[4706]: E1127 07:10:42.777023 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:42 crc kubenswrapper[4706]: E1127 07:10:42.777306 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:43 crc kubenswrapper[4706]: I1127 07:10:43.776295 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:43 crc kubenswrapper[4706]: E1127 07:10:43.776450 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:44 crc kubenswrapper[4706]: I1127 07:10:44.776769 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:44 crc kubenswrapper[4706]: E1127 07:10:44.777054 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:44 crc kubenswrapper[4706]: I1127 07:10:44.777396 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:44 crc kubenswrapper[4706]: E1127 07:10:44.777538 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:44 crc kubenswrapper[4706]: I1127 07:10:44.777722 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:44 crc kubenswrapper[4706]: E1127 07:10:44.777806 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:45 crc kubenswrapper[4706]: I1127 07:10:45.775732 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:45 crc kubenswrapper[4706]: E1127 07:10:45.775920 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:46 crc kubenswrapper[4706]: I1127 07:10:46.776760 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:46 crc kubenswrapper[4706]: I1127 07:10:46.776834 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:46 crc kubenswrapper[4706]: I1127 07:10:46.776871 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:46 crc kubenswrapper[4706]: E1127 07:10:46.777999 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:46 crc kubenswrapper[4706]: E1127 07:10:46.778088 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:46 crc kubenswrapper[4706]: E1127 07:10:46.778377 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:46 crc kubenswrapper[4706]: I1127 07:10:46.778638 4706 scope.go:117] "RemoveContainer" containerID="5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6" Nov 27 07:10:46 crc kubenswrapper[4706]: E1127 07:10:46.895443 4706 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 07:10:47 crc kubenswrapper[4706]: I1127 07:10:47.380693 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/1.log" Nov 27 07:10:47 crc kubenswrapper[4706]: I1127 07:10:47.381346 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerStarted","Data":"408dbfa8d92fabe6661bc2452e5738f59d4453f29fd4329bb8b6037bc5acd519"} Nov 27 07:10:47 crc kubenswrapper[4706]: I1127 07:10:47.776548 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:47 crc kubenswrapper[4706]: E1127 07:10:47.776713 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:48 crc kubenswrapper[4706]: I1127 07:10:48.776092 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:48 crc kubenswrapper[4706]: I1127 07:10:48.776097 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:48 crc kubenswrapper[4706]: I1127 07:10:48.776268 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:48 crc kubenswrapper[4706]: E1127 07:10:48.776353 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:48 crc kubenswrapper[4706]: E1127 07:10:48.776563 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:48 crc kubenswrapper[4706]: E1127 07:10:48.776677 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:49 crc kubenswrapper[4706]: I1127 07:10:49.776742 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:49 crc kubenswrapper[4706]: E1127 07:10:49.777808 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:50 crc kubenswrapper[4706]: I1127 07:10:50.776920 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:50 crc kubenswrapper[4706]: I1127 07:10:50.776944 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:50 crc kubenswrapper[4706]: I1127 07:10:50.777094 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:50 crc kubenswrapper[4706]: E1127 07:10:50.777409 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 07:10:50 crc kubenswrapper[4706]: E1127 07:10:50.777528 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 07:10:50 crc kubenswrapper[4706]: E1127 07:10:50.777655 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 07:10:51 crc kubenswrapper[4706]: I1127 07:10:51.776661 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:51 crc kubenswrapper[4706]: E1127 07:10:51.776897 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zxp2h" podUID="ab5bf95b-5c0d-4804-bc97-43b79f6f2fea" Nov 27 07:10:52 crc kubenswrapper[4706]: I1127 07:10:52.776809 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:10:52 crc kubenswrapper[4706]: I1127 07:10:52.776951 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:10:52 crc kubenswrapper[4706]: I1127 07:10:52.777253 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:10:52 crc kubenswrapper[4706]: I1127 07:10:52.780160 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 07:10:52 crc kubenswrapper[4706]: I1127 07:10:52.780522 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 07:10:52 crc kubenswrapper[4706]: I1127 07:10:52.781813 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 07:10:52 crc kubenswrapper[4706]: I1127 07:10:52.781895 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 07:10:53 crc kubenswrapper[4706]: I1127 07:10:53.776860 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:10:53 crc kubenswrapper[4706]: I1127 07:10:53.780150 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 07:10:53 crc kubenswrapper[4706]: I1127 07:10:53.780368 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.926674 4706 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.967575 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k"] Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.967944 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.971184 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.972337 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.973065 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.973309 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.973457 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.973785 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.975964 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ghm8f"] Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.976457 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.978104 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl"] Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.978722 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.979362 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j"] Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.979716 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.980875 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6"] Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.981204 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:55 crc kubenswrapper[4706]: I1127 07:10:55.981790 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-n2f45"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:55.996152 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.013521 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.013858 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.013956 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.014262 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.014420 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.014541 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.014826 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.014976 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.014445 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-snqv9"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.015748 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-dh98x"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.016308 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.016641 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.016787 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lbldb"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.017273 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.017361 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.017502 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.017664 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dh98x" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.017732 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.017930 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.017955 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018474 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018584 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018618 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018722 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018731 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018782 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018836 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.018784 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.021112 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.022946 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.023347 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.023560 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.023751 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.024516 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.024594 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.024863 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.025023 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.027727 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r6h9j"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.028505 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.028790 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-nhj5d"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.029336 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.029468 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.029998 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.030500 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.030919 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f39b7c35-28bf-444f-a699-8607c48b6b62-serving-cert\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.030965 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-service-ca-bundle\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.030985 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxjf8\" (UniqueName: \"kubernetes.io/projected/f39b7c35-28bf-444f-a699-8607c48b6b62-kube-api-access-dxjf8\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031039 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-client-ca\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031071 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0558cfe-6d59-4897-9a10-595f42734ca6-serving-cert\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031074 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031093 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-config\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031128 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-config\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031143 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031170 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k299h\" (UniqueName: \"kubernetes.io/projected/a0558cfe-6d59-4897-9a10-595f42734ca6-kube-api-access-k299h\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031231 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031511 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031678 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.031904 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032121 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032169 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032208 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032090 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wt6br"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032356 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032421 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032736 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032825 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032890 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.032963 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.036349 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.038776 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fqglc"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.039245 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.044768 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.044833 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.044963 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.044975 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.045262 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mtl4k"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.045393 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.045407 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.045561 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.045581 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.046055 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.048018 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.048497 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.048771 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.049494 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.050317 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.050995 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.052349 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2gpf6"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.085393 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.086106 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.086462 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.086649 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.087072 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.088343 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.089506 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.089699 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.089921 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.096851 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.096952 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.099484 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.100003 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.100018 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.100539 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-8bwv8"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.104047 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.104607 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.104771 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.105148 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.105191 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.108132 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.108372 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.108505 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.108640 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.108764 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.108856 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.108943 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.109125 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.110918 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.111387 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.111602 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.111718 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.112657 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.113082 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.114340 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.115821 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.117135 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.121524 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.122174 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.122365 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.122511 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.122659 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.122700 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.122751 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.122886 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.123107 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.123139 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.123314 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.123358 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.123464 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.123600 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.123765 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.125276 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.126068 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.126372 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.126545 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.127571 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.128097 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.128397 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.128938 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.129527 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.130029 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.130504 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.132143 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.144693 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.133351 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.138235 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146394 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/41f449f3-95c5-4193-81a2-ade651d73501-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146472 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-config\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146501 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3939cbca-25db-434c-8527-ec3bc8e1aa85-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146527 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-image-import-ca\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146551 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f449f3-95c5-4193-81a2-ade651d73501-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146578 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e36cad-dbdd-458c-9862-7131d620ba34-console-serving-cert\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146599 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8e36cad-dbdd-458c-9862-7131d620ba34-console-oauth-config\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146623 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-serving-cert\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146647 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6drrq\" (UniqueName: \"kubernetes.io/projected/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-kube-api-access-6drrq\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146676 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v48f\" (UniqueName: \"kubernetes.io/projected/1fbc1beb-c412-4ca8-ae88-8c73e17803b9-kube-api-access-5v48f\") pod \"cluster-samples-operator-665b6dd947-8ndx6\" (UID: \"1fbc1beb-c412-4ca8-ae88-8c73e17803b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146710 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-etcd-serving-ca\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146734 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-serving-cert\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146763 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbkhn\" (UniqueName: \"kubernetes.io/projected/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-kube-api-access-rbkhn\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146797 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-service-ca-bundle\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146822 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwj8d\" (UniqueName: \"kubernetes.io/projected/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-kube-api-access-xwj8d\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146847 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6b3e5d2-bd7d-402d-af78-f685211e7f8e-metrics-tls\") pod \"dns-operator-744455d44c-n2f45\" (UID: \"a6b3e5d2-bd7d-402d-af78-f685211e7f8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146893 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxjf8\" (UniqueName: \"kubernetes.io/projected/f39b7c35-28bf-444f-a699-8607c48b6b62-kube-api-access-dxjf8\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146916 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-dir\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146939 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-config\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.146963 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-service-ca\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147007 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-service-ca\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147053 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-client-ca\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147075 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147101 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-config\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147127 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-oauth-serving-cert\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147157 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9h6d\" (UniqueName: \"kubernetes.io/projected/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-kube-api-access-r9h6d\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147186 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147234 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147264 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-serving-cert\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147293 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl2lb\" (UniqueName: \"kubernetes.io/projected/7f836197-8a96-4af0-b3bc-eb46a667b71c-kube-api-access-hl2lb\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147325 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cfzh\" (UniqueName: \"kubernetes.io/projected/d8e36cad-dbdd-458c-9862-7131d620ba34-kube-api-access-9cfzh\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147354 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k299h\" (UniqueName: \"kubernetes.io/projected/a0558cfe-6d59-4897-9a10-595f42734ca6-kube-api-access-k299h\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147379 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-policies\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147434 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147467 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb4d0f97-a44e-4323-b52d-c060c71684ff-config\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147496 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnltd\" (UniqueName: \"kubernetes.io/projected/0b4e82db-e890-4dc6-a6f0-441d78878459-kube-api-access-bnltd\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147520 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-etcd-client\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147544 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-machine-approver-tls\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147566 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzp7g\" (UniqueName: \"kubernetes.io/projected/a6b3e5d2-bd7d-402d-af78-f685211e7f8e-kube-api-access-xzp7g\") pod \"dns-operator-744455d44c-n2f45\" (UID: \"a6b3e5d2-bd7d-402d-af78-f685211e7f8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147592 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b4e82db-e890-4dc6-a6f0-441d78878459-config\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147616 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7950dc1-2f65-4f07-9a5c-3f847effd16b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147639 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7950dc1-2f65-4f07-9a5c-3f847effd16b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147657 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-trusted-ca\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147672 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147680 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-config\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.148342 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-auth-proxy-config\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.148381 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvqph\" (UniqueName: \"kubernetes.io/projected/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-kube-api-access-dvqph\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.148664 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-config\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.148675 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-service-ca-bundle\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.148692 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-encryption-config\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.147293 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154167 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk9rz\" (UniqueName: \"kubernetes.io/projected/f7950dc1-2f65-4f07-9a5c-3f847effd16b-kube-api-access-hk9rz\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154255 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f39b7c35-28bf-444f-a699-8607c48b6b62-serving-cert\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154699 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154724 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmhsn\" (UniqueName: \"kubernetes.io/projected/3939cbca-25db-434c-8527-ec3bc8e1aa85-kube-api-access-zmhsn\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154828 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0b4e82db-e890-4dc6-a6f0-441d78878459-images\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154864 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154887 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154906 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-console-config\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154928 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154946 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3939cbca-25db-434c-8527-ec3bc8e1aa85-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154965 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.154981 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb4d0f97-a44e-4323-b52d-c060c71684ff-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155009 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155027 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w68w6\" (UniqueName: \"kubernetes.io/projected/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-kube-api-access-w68w6\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155046 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-client\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155079 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1fbc1beb-c412-4ca8-ae88-8c73e17803b9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8ndx6\" (UID: \"1fbc1beb-c412-4ca8-ae88-8c73e17803b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155099 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-serving-cert\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155125 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-client-ca\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155144 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-trusted-ca-bundle\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155166 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-audit\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155210 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0558cfe-6d59-4897-9a10-595f42734ca6-serving-cert\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155244 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-config\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155267 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-serving-cert\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155284 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f836197-8a96-4af0-b3bc-eb46a667b71c-audit-dir\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155301 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p27w\" (UniqueName: \"kubernetes.io/projected/41f449f3-95c5-4193-81a2-ade651d73501-kube-api-access-4p27w\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155321 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-config\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155342 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0b4e82db-e890-4dc6-a6f0-441d78878459-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155358 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41f449f3-95c5-4193-81a2-ade651d73501-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155374 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb4d0f97-a44e-4323-b52d-c060c71684ff-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155376 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155625 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.156291 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.155392 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4xnk\" (UniqueName: \"kubernetes.io/projected/094c9ba3-fb86-48e5-93e0-f5f1f55cc12e-kube-api-access-f4xnk\") pod \"downloads-7954f5f757-dh98x\" (UID: \"094c9ba3-fb86-48e5-93e0-f5f1f55cc12e\") " pod="openshift-console/downloads-7954f5f757-dh98x" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.156707 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.156918 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157008 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rm5f8"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157096 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f39b7c35-28bf-444f-a699-8607c48b6b62-config\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157116 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157249 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-client-ca\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157410 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157445 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-available-featuregates\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157489 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157520 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f836197-8a96-4af0-b3bc-eb46a667b71c-node-pullsecrets\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157541 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-ca\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.157737 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.158946 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-config\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.159258 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.161296 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.161620 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f39b7c35-28bf-444f-a699-8607c48b6b62-serving-cert\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.161937 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0558cfe-6d59-4897-9a10-595f42734ca6-serving-cert\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.162304 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.164303 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.166370 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.167026 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.168231 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-np86m"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.168298 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.168924 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.169538 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jgpht"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.169643 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.170303 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.170811 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.172842 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r6h9j"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.174178 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.176650 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ghm8f"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.180170 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-n2f45"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.182614 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rns7p"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.183512 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.183631 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.184633 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.185759 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.187137 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.188488 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dh98x"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.190132 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.190681 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.193507 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.195357 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.202470 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.205172 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mtl4k"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.205197 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lbldb"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.206102 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wt6br"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.207504 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fqglc"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.208310 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-nhj5d"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.209301 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.210309 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-snqv9"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.211247 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.211404 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.214165 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.214195 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.215312 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-v9mq5"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.216146 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.216415 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.217780 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2gpf6"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.219032 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.224733 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.227432 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.228739 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.229971 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.230315 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jgpht"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.231378 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.232610 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.241844 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rns7p"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.245882 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v9mq5"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.246277 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-np86m"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.247360 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.249746 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rm5f8"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.251180 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8jrds"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.252073 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.252945 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-wcx4x"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.253100 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.254348 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8jrds"] Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.254454 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258192 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-audit\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258245 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258283 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-serving-cert\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258303 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f836197-8a96-4af0-b3bc-eb46a667b71c-audit-dir\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258320 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41f449f3-95c5-4193-81a2-ade651d73501-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258340 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4xnk\" (UniqueName: \"kubernetes.io/projected/094c9ba3-fb86-48e5-93e0-f5f1f55cc12e-kube-api-access-f4xnk\") pod \"downloads-7954f5f757-dh98x\" (UID: \"094c9ba3-fb86-48e5-93e0-f5f1f55cc12e\") " pod="openshift-console/downloads-7954f5f757-dh98x" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258360 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258376 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f836197-8a96-4af0-b3bc-eb46a667b71c-node-pullsecrets\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258393 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-ca\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258410 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258429 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258447 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-images\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258580 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3939cbca-25db-434c-8527-ec3bc8e1aa85-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258645 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8e36cad-dbdd-458c-9862-7131d620ba34-console-oauth-config\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258671 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-serving-cert\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258691 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6drrq\" (UniqueName: \"kubernetes.io/projected/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-kube-api-access-6drrq\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258713 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v48f\" (UniqueName: \"kubernetes.io/projected/1fbc1beb-c412-4ca8-ae88-8c73e17803b9-kube-api-access-5v48f\") pod \"cluster-samples-operator-665b6dd947-8ndx6\" (UID: \"1fbc1beb-c412-4ca8-ae88-8c73e17803b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258735 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-etcd-serving-ca\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258755 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbkhn\" (UniqueName: \"kubernetes.io/projected/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-kube-api-access-rbkhn\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258776 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a9235278-ebf9-42f9-bfce-945f5453f26d-srv-cert\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258798 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwj8d\" (UniqueName: \"kubernetes.io/projected/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-kube-api-access-xwj8d\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6b3e5d2-bd7d-402d-af78-f685211e7f8e-metrics-tls\") pod \"dns-operator-744455d44c-n2f45\" (UID: \"a6b3e5d2-bd7d-402d-af78-f685211e7f8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258844 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d91244b2-5d79-4540-a2ae-41ae33512da4-proxy-tls\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258866 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-service-ca\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258883 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258905 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtcv7\" (UniqueName: \"kubernetes.io/projected/c7b9d8c1-2a4b-44f3-8d20-649b1c893018-kube-api-access-gtcv7\") pod \"migrator-59844c95c7-bmvl6\" (UID: \"c7b9d8c1-2a4b-44f3-8d20-649b1c893018\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258942 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-config\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258970 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-service-ca\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258963 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f836197-8a96-4af0-b3bc-eb46a667b71c-node-pullsecrets\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.258993 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94bada3b-c6c2-4e88-813c-becf573abbe2-cert\") pod \"ingress-canary-rns7p\" (UID: \"94bada3b-c6c2-4e88-813c-becf573abbe2\") " pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259027 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f836197-8a96-4af0-b3bc-eb46a667b71c-audit-dir\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259041 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-audit-policies\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259088 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df22a91c-9a68-4b08-972f-397c91bbc2d9-audit-dir\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259121 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/417efb56-71f3-4d60-bfb4-c1b11448f3be-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259146 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-serving-cert\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259187 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75bd3901-cc36-42ee-a1b5-78fd448ced38-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259209 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d8j2\" (UniqueName: \"kubernetes.io/projected/a9235278-ebf9-42f9-bfce-945f5453f26d-kube-api-access-8d8j2\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259256 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dd986c7f-e3ae-4d24-8815-c38d546168e7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c878d\" (UID: \"dd986c7f-e3ae-4d24-8815-c38d546168e7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259281 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-policies\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259287 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3939cbca-25db-434c-8527-ec3bc8e1aa85-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259301 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259323 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/417efb56-71f3-4d60-bfb4-c1b11448f3be-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259393 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259425 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnltd\" (UniqueName: \"kubernetes.io/projected/0b4e82db-e890-4dc6-a6f0-441d78878459-kube-api-access-bnltd\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259451 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-etcd-client\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259468 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzp7g\" (UniqueName: \"kubernetes.io/projected/a6b3e5d2-bd7d-402d-af78-f685211e7f8e-kube-api-access-xzp7g\") pod \"dns-operator-744455d44c-n2f45\" (UID: \"a6b3e5d2-bd7d-402d-af78-f685211e7f8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259487 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-config\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259508 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-serving-cert\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259593 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b4e82db-e890-4dc6-a6f0-441d78878459-config\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259615 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7950dc1-2f65-4f07-9a5c-3f847effd16b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259635 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-auth-proxy-config\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259658 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w49c\" (UniqueName: \"kubernetes.io/projected/d91244b2-5d79-4540-a2ae-41ae33512da4-kube-api-access-4w49c\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259678 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td7gw\" (UniqueName: \"kubernetes.io/projected/dd986c7f-e3ae-4d24-8815-c38d546168e7-kube-api-access-td7gw\") pod \"control-plane-machine-set-operator-78cbb6b69f-c878d\" (UID: \"dd986c7f-e3ae-4d24-8815-c38d546168e7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259702 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-config\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259729 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.259747 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-console-config\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.260118 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-audit\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.260195 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-service-ca\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.261640 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-etcd-serving-ca\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.261748 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262061 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-policies\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262102 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-config\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262273 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-config\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262302 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-ca\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262668 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6b3e5d2-bd7d-402d-af78-f685211e7f8e-metrics-tls\") pod \"dns-operator-744455d44c-n2f45\" (UID: \"a6b3e5d2-bd7d-402d-af78-f685211e7f8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262890 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-service-ca\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-config\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262872 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.262928 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0b4e82db-e890-4dc6-a6f0-441d78878459-images\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w68w6\" (UniqueName: \"kubernetes.io/projected/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-kube-api-access-w68w6\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263090 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-console-config\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263267 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1fbc1beb-c412-4ca8-ae88-8c73e17803b9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8ndx6\" (UID: \"1fbc1beb-c412-4ca8-ae88-8c73e17803b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263318 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-trusted-ca-bundle\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263327 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-auth-proxy-config\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263361 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77b22\" (UniqueName: \"kubernetes.io/projected/94bada3b-c6c2-4e88-813c-becf573abbe2-kube-api-access-77b22\") pod \"ingress-canary-rns7p\" (UID: \"94bada3b-c6c2-4e88-813c-becf573abbe2\") " pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263385 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/75bd3901-cc36-42ee-a1b5-78fd448ced38-metrics-tls\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263415 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj4n9\" (UniqueName: \"kubernetes.io/projected/ba960b0a-5a50-4165-8572-0062d1ebff58-kube-api-access-cj4n9\") pod \"multus-admission-controller-857f4d67dd-rm5f8\" (UID: \"ba960b0a-5a50-4165-8572-0062d1ebff58\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263455 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3aaba837-66f8-4b9e-a3d8-883406929734-webhook-cert\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263474 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-957vs\" (UniqueName: \"kubernetes.io/projected/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-kube-api-access-957vs\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-config\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263528 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263535 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7950dc1-2f65-4f07-9a5c-3f847effd16b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263548 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p27w\" (UniqueName: \"kubernetes.io/projected/41f449f3-95c5-4193-81a2-ade651d73501-kube-api-access-4p27w\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263575 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91244b2-5d79-4540-a2ae-41ae33512da4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263632 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0b4e82db-e890-4dc6-a6f0-441d78878459-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263656 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb4d0f97-a44e-4323-b52d-c060c71684ff-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263680 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263705 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-available-featuregates\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263729 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-proxy-tls\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.263886 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0b4e82db-e890-4dc6-a6f0-441d78878459-images\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264309 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-config\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264367 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b4e82db-e890-4dc6-a6f0-441d78878459-config\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264449 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57pcz\" (UniqueName: \"kubernetes.io/projected/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-kube-api-access-57pcz\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264572 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264630 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/41f449f3-95c5-4193-81a2-ade651d73501-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264655 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264688 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-key\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264713 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264894 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-trusted-ca-bundle\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.264926 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-serving-cert\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265005 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-config\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265022 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-available-featuregates\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265045 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-image-import-ca\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265120 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-etcd-client\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265187 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f449f3-95c5-4193-81a2-ade651d73501-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265264 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb7b9\" (UniqueName: \"kubernetes.io/projected/75bd3901-cc36-42ee-a1b5-78fd448ced38-kube-api-access-kb7b9\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265415 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e36cad-dbdd-458c-9862-7131d620ba34-console-serving-cert\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2bg9\" (UniqueName: \"kubernetes.io/projected/a709c580-4d9f-457d-a6cf-67a2831f4dcd-kube-api-access-h2bg9\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265630 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265706 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-serving-cert\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265645 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8e36cad-dbdd-458c-9862-7131d620ba34-console-oauth-config\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265645 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265841 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2snhf\" (UniqueName: \"kubernetes.io/projected/a4dcfcf5-d873-4adb-868a-58e1630b816a-kube-api-access-2snhf\") pod \"package-server-manager-789f6589d5-gdp66\" (UID: \"a4dcfcf5-d873-4adb-868a-58e1630b816a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.265903 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-config\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-image-import-ca\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266121 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-encryption-config\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266215 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f449f3-95c5-4193-81a2-ade651d73501-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266258 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-config\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266471 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-dir\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266604 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlwvh\" (UniqueName: \"kubernetes.io/projected/df22a91c-9a68-4b08-972f-397c91bbc2d9-kube-api-access-hlwvh\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266742 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99cqs\" (UniqueName: \"kubernetes.io/projected/417efb56-71f3-4d60-bfb4-c1b11448f3be-kube-api-access-99cqs\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266868 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba960b0a-5a50-4165-8572-0062d1ebff58-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rm5f8\" (UID: \"ba960b0a-5a50-4165-8572-0062d1ebff58\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266964 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3aaba837-66f8-4b9e-a3d8-883406929734-tmpfs\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267048 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqxn9\" (UniqueName: \"kubernetes.io/projected/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-kube-api-access-qqxn9\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267151 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267252 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266473 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-serving-cert\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.266508 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-dir\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267157 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267461 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a9235278-ebf9-42f9-bfce-945f5453f26d-profile-collector-cert\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-oauth-serving-cert\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267628 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9h6d\" (UniqueName: \"kubernetes.io/projected/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-kube-api-access-r9h6d\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267708 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a709c580-4d9f-457d-a6cf-67a2831f4dcd-secret-volume\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267762 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e36cad-dbdd-458c-9862-7131d620ba34-console-serving-cert\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267874 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/04db01d1-964d-49eb-9b17-4f81da5b31f9-srv-cert\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267960 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268046 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268126 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl2lb\" (UniqueName: \"kubernetes.io/projected/7f836197-8a96-4af0-b3bc-eb46a667b71c-kube-api-access-hl2lb\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268202 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0b4e82db-e890-4dc6-a6f0-441d78878459-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268202 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cfzh\" (UniqueName: \"kubernetes.io/projected/d8e36cad-dbdd-458c-9862-7131d620ba34-kube-api-access-9cfzh\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.267927 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1fbc1beb-c412-4ca8-ae88-8c73e17803b9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8ndx6\" (UID: \"1fbc1beb-c412-4ca8-ae88-8c73e17803b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268280 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3aaba837-66f8-4b9e-a3d8-883406929734-apiservice-cert\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268306 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb4d0f97-a44e-4323-b52d-c060c71684ff-config\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-machine-approver-tls\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268356 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn7kx\" (UniqueName: \"kubernetes.io/projected/04db01d1-964d-49eb-9b17-4f81da5b31f9-kube-api-access-bn7kx\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268382 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7950dc1-2f65-4f07-9a5c-3f847effd16b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268401 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-trusted-ca\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268419 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-etcd-client\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268444 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvqph\" (UniqueName: \"kubernetes.io/projected/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-kube-api-access-dvqph\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268463 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-encryption-config\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268480 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk9rz\" (UniqueName: \"kubernetes.io/projected/f7950dc1-2f65-4f07-9a5c-3f847effd16b-kube-api-access-hk9rz\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268501 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268520 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmhsn\" (UniqueName: \"kubernetes.io/projected/3939cbca-25db-434c-8527-ec3bc8e1aa85-kube-api-access-zmhsn\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268546 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268590 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb4d0f97-a44e-4323-b52d-c060c71684ff-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268611 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/04db01d1-964d-49eb-9b17-4f81da5b31f9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268634 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268654 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3939cbca-25db-434c-8527-ec3bc8e1aa85-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268674 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268692 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-client\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268713 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-cabundle\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268738 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-serving-cert\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268759 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4dcfcf5-d873-4adb-868a-58e1630b816a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gdp66\" (UID: \"a4dcfcf5-d873-4adb-868a-58e1630b816a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268781 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plh8h\" (UniqueName: \"kubernetes.io/projected/3aaba837-66f8-4b9e-a3d8-883406929734-kube-api-access-plh8h\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268803 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-client-ca\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268824 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75bd3901-cc36-42ee-a1b5-78fd448ced38-trusted-ca\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268867 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb4d0f97-a44e-4323-b52d-c060c71684ff-config\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.268899 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/41f449f3-95c5-4193-81a2-ade651d73501-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.269029 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-serving-cert\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.269431 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8e36cad-dbdd-458c-9862-7131d620ba34-oauth-serving-cert\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.270826 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.270902 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-trusted-ca\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.271127 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-serving-cert\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.271246 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.271635 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f836197-8a96-4af0-b3bc-eb46a667b71c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.272081 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-client-ca\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.273163 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.273561 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.273687 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-serving-cert\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.273703 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.273690 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb4d0f97-a44e-4323-b52d-c060c71684ff-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.274294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7f836197-8a96-4af0-b3bc-eb46a667b71c-encryption-config\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.274367 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7950dc1-2f65-4f07-9a5c-3f847effd16b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.274429 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.274974 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.275481 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-machine-approver-tls\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.275626 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3939cbca-25db-434c-8527-ec3bc8e1aa85-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.276797 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-etcd-client\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.290877 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.315993 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.330186 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.350123 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369714 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w49c\" (UniqueName: \"kubernetes.io/projected/d91244b2-5d79-4540-a2ae-41ae33512da4-kube-api-access-4w49c\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369754 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td7gw\" (UniqueName: \"kubernetes.io/projected/dd986c7f-e3ae-4d24-8815-c38d546168e7-kube-api-access-td7gw\") pod \"control-plane-machine-set-operator-78cbb6b69f-c878d\" (UID: \"dd986c7f-e3ae-4d24-8815-c38d546168e7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369788 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj4n9\" (UniqueName: \"kubernetes.io/projected/ba960b0a-5a50-4165-8572-0062d1ebff58-kube-api-access-cj4n9\") pod \"multus-admission-controller-857f4d67dd-rm5f8\" (UID: \"ba960b0a-5a50-4165-8572-0062d1ebff58\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369808 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3aaba837-66f8-4b9e-a3d8-883406929734-webhook-cert\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369826 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-957vs\" (UniqueName: \"kubernetes.io/projected/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-kube-api-access-957vs\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369844 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77b22\" (UniqueName: \"kubernetes.io/projected/94bada3b-c6c2-4e88-813c-becf573abbe2-kube-api-access-77b22\") pod \"ingress-canary-rns7p\" (UID: \"94bada3b-c6c2-4e88-813c-becf573abbe2\") " pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369860 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/75bd3901-cc36-42ee-a1b5-78fd448ced38-metrics-tls\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369884 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369908 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91244b2-5d79-4540-a2ae-41ae33512da4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369927 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-proxy-tls\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369947 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57pcz\" (UniqueName: \"kubernetes.io/projected/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-kube-api-access-57pcz\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369964 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369978 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-key\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.369992 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370010 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2bg9\" (UniqueName: \"kubernetes.io/projected/a709c580-4d9f-457d-a6cf-67a2831f4dcd-kube-api-access-h2bg9\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370027 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb7b9\" (UniqueName: \"kubernetes.io/projected/75bd3901-cc36-42ee-a1b5-78fd448ced38-kube-api-access-kb7b9\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370045 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2snhf\" (UniqueName: \"kubernetes.io/projected/a4dcfcf5-d873-4adb-868a-58e1630b816a-kube-api-access-2snhf\") pod \"package-server-manager-789f6589d5-gdp66\" (UID: \"a4dcfcf5-d873-4adb-868a-58e1630b816a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370061 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-encryption-config\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370077 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-config\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370094 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlwvh\" (UniqueName: \"kubernetes.io/projected/df22a91c-9a68-4b08-972f-397c91bbc2d9-kube-api-access-hlwvh\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370110 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3aaba837-66f8-4b9e-a3d8-883406929734-tmpfs\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370128 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqxn9\" (UniqueName: \"kubernetes.io/projected/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-kube-api-access-qqxn9\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370147 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99cqs\" (UniqueName: \"kubernetes.io/projected/417efb56-71f3-4d60-bfb4-c1b11448f3be-kube-api-access-99cqs\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370165 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba960b0a-5a50-4165-8572-0062d1ebff58-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rm5f8\" (UID: \"ba960b0a-5a50-4165-8572-0062d1ebff58\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370190 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370206 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a9235278-ebf9-42f9-bfce-945f5453f26d-profile-collector-cert\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370248 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a709c580-4d9f-457d-a6cf-67a2831f4dcd-secret-volume\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370279 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/04db01d1-964d-49eb-9b17-4f81da5b31f9-srv-cert\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370295 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370320 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3aaba837-66f8-4b9e-a3d8-883406929734-apiservice-cert\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370337 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn7kx\" (UniqueName: \"kubernetes.io/projected/04db01d1-964d-49eb-9b17-4f81da5b31f9-kube-api-access-bn7kx\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370354 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-etcd-client\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370391 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/04db01d1-964d-49eb-9b17-4f81da5b31f9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370415 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-cabundle\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370434 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4dcfcf5-d873-4adb-868a-58e1630b816a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gdp66\" (UID: \"a4dcfcf5-d873-4adb-868a-58e1630b816a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370452 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plh8h\" (UniqueName: \"kubernetes.io/projected/3aaba837-66f8-4b9e-a3d8-883406929734-kube-api-access-plh8h\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75bd3901-cc36-42ee-a1b5-78fd448ced38-trusted-ca\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370547 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370565 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-images\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370599 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a9235278-ebf9-42f9-bfce-945f5453f26d-srv-cert\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370630 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d91244b2-5d79-4540-a2ae-41ae33512da4-proxy-tls\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370650 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtcv7\" (UniqueName: \"kubernetes.io/projected/c7b9d8c1-2a4b-44f3-8d20-649b1c893018-kube-api-access-gtcv7\") pod \"migrator-59844c95c7-bmvl6\" (UID: \"c7b9d8c1-2a4b-44f3-8d20-649b1c893018\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370685 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370703 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94bada3b-c6c2-4e88-813c-becf573abbe2-cert\") pod \"ingress-canary-rns7p\" (UID: \"94bada3b-c6c2-4e88-813c-becf573abbe2\") " pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370723 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-audit-policies\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370727 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3aaba837-66f8-4b9e-a3d8-883406929734-tmpfs\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370756 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df22a91c-9a68-4b08-972f-397c91bbc2d9-audit-dir\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370775 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/417efb56-71f3-4d60-bfb4-c1b11448f3be-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370791 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75bd3901-cc36-42ee-a1b5-78fd448ced38-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370811 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d8j2\" (UniqueName: \"kubernetes.io/projected/a9235278-ebf9-42f9-bfce-945f5453f26d-kube-api-access-8d8j2\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370830 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dd986c7f-e3ae-4d24-8815-c38d546168e7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c878d\" (UID: \"dd986c7f-e3ae-4d24-8815-c38d546168e7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370852 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/417efb56-71f3-4d60-bfb4-c1b11448f3be-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.370883 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-serving-cert\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.371046 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df22a91c-9a68-4b08-972f-397c91bbc2d9-audit-dir\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.371198 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91244b2-5d79-4540-a2ae-41ae33512da4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.371245 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.371752 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75bd3901-cc36-42ee-a1b5-78fd448ced38-trusted-ca\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.372669 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.374214 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/75bd3901-cc36-42ee-a1b5-78fd448ced38-metrics-tls\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.391181 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.412971 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.430067 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.450938 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.471002 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.491099 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.510936 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.530968 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.550826 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.570747 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.590659 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.612330 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.629713 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.651173 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.670828 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.686319 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.690099 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.690736 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-config\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.730651 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.750368 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.770731 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.791133 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.810465 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.812119 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-audit-policies\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.831026 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.835192 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/417efb56-71f3-4d60-bfb4-c1b11448f3be-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.851024 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.870413 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.890391 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.892436 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/417efb56-71f3-4d60-bfb4-c1b11448f3be-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.910739 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.929972 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.933355 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-etcd-client\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.950624 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.956357 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-serving-cert\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.971997 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.984244 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/df22a91c-9a68-4b08-972f-397c91bbc2d9-encryption-config\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:56 crc kubenswrapper[4706]: I1127 07:10:56.992138 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.001256 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.011681 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.021291 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df22a91c-9a68-4b08-972f-397c91bbc2d9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.031847 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.051098 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.070327 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.091720 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.110834 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.126377 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3aaba837-66f8-4b9e-a3d8-883406929734-apiservice-cert\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.126446 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3aaba837-66f8-4b9e-a3d8-883406929734-webhook-cert\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.131904 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.149128 4706 request.go:700] Waited for 1.010381932s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpprof-cert&limit=500&resourceVersion=0 Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.156053 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.165886 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a709c580-4d9f-457d-a6cf-67a2831f4dcd-secret-volume\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.166564 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/04db01d1-964d-49eb-9b17-4f81da5b31f9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.167422 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a9235278-ebf9-42f9-bfce-945f5453f26d-profile-collector-cert\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.171748 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.185933 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a9235278-ebf9-42f9-bfce-945f5453f26d-srv-cert\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.211190 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.212911 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxjf8\" (UniqueName: \"kubernetes.io/projected/f39b7c35-28bf-444f-a699-8607c48b6b62-kube-api-access-dxjf8\") pod \"authentication-operator-69f744f599-ghm8f\" (UID: \"f39b7c35-28bf-444f-a699-8607c48b6b62\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.226055 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4dcfcf5-d873-4adb-868a-58e1630b816a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gdp66\" (UID: \"a4dcfcf5-d873-4adb-868a-58e1630b816a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.231009 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.231378 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.239098 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d91244b2-5d79-4540-a2ae-41ae33512da4-proxy-tls\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.270481 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k299h\" (UniqueName: \"kubernetes.io/projected/a0558cfe-6d59-4897-9a10-595f42734ca6-kube-api-access-k299h\") pod \"route-controller-manager-6576b87f9c-q7d5k\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.273042 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.281789 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-images\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.290181 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.310700 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.332255 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.359784 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/04db01d1-964d-49eb-9b17-4f81da5b31f9-srv-cert\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.361282 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.369415 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-proxy-tls\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.370040 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370453 4706 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370557 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-key podName:7d0da76e-b483-4a6a-a1b3-f4c64a29ddca nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.8705281 +0000 UTC m=+141.760118940 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-key") pod "service-ca-9c57cc56f-np86m" (UID: "7d0da76e-b483-4a6a-a1b3-f4c64a29ddca") : failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370644 4706 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370677 4706 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370771 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba960b0a-5a50-4165-8572-0062d1ebff58-webhook-certs podName:ba960b0a-5a50-4165-8572-0062d1ebff58 nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.870743517 +0000 UTC m=+141.760334327 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ba960b0a-5a50-4165-8572-0062d1ebff58-webhook-certs") pod "multus-admission-controller-857f4d67dd-rm5f8" (UID: "ba960b0a-5a50-4165-8572-0062d1ebff58") : failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370796 4706 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370814 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics podName:e22ae6d0-dbdf-4b69-84f0-0643266a7dff nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.87080392 +0000 UTC m=+141.760394730 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics") pod "marketplace-operator-79b997595-jgpht" (UID: "e22ae6d0-dbdf-4b69-84f0-0643266a7dff") : failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370772 4706 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370827 4706 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370868 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume podName:a709c580-4d9f-457d-a6cf-67a2831f4dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.870847931 +0000 UTC m=+141.760438741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume") pod "collect-profiles-29403780-gsqzn" (UID: "a709c580-4d9f-457d-a6cf-67a2831f4dcd") : failed to sync configmap cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370896 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-cabundle podName:7d0da76e-b483-4a6a-a1b3-f4c64a29ddca nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.870877892 +0000 UTC m=+141.760468742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-cabundle") pod "service-ca-9c57cc56f-np86m" (UID: "7d0da76e-b483-4a6a-a1b3-f4c64a29ddca") : failed to sync configmap cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.370930 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca podName:e22ae6d0-dbdf-4b69-84f0-0643266a7dff nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.870913444 +0000 UTC m=+141.760504294 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca") pod "marketplace-operator-79b997595-jgpht" (UID: "e22ae6d0-dbdf-4b69-84f0-0643266a7dff") : failed to sync configmap cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.371877 4706 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.371947 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dd986c7f-e3ae-4d24-8815-c38d546168e7-control-plane-machine-set-operator-tls podName:dd986c7f-e3ae-4d24-8815-c38d546168e7 nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.87193398 +0000 UTC m=+141.761524780 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/dd986c7f-e3ae-4d24-8815-c38d546168e7-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-c878d" (UID: "dd986c7f-e3ae-4d24-8815-c38d546168e7") : failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.372591 4706 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: E1127 07:10:57.372680 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/94bada3b-c6c2-4e88-813c-becf573abbe2-cert podName:94bada3b-c6c2-4e88-813c-becf573abbe2 nodeName:}" failed. No retries permitted until 2025-11-27 07:10:57.872644626 +0000 UTC m=+141.762235616 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/94bada3b-c6c2-4e88-813c-becf573abbe2-cert") pod "ingress-canary-rns7p" (UID: "94bada3b-c6c2-4e88-813c-becf573abbe2") : failed to sync secret cache: timed out waiting for the condition Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.391881 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.415175 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.431278 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.451005 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.472515 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.491236 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.509982 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.523902 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.526988 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ghm8f"] Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.531665 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 07:10:57 crc kubenswrapper[4706]: W1127 07:10:57.544697 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf39b7c35_28bf_444f_a699_8607c48b6b62.slice/crio-8bbbcce9e1f4b38559aeca4b625686dfa611638475c6929685846a3d119b8486 WatchSource:0}: Error finding container 8bbbcce9e1f4b38559aeca4b625686dfa611638475c6929685846a3d119b8486: Status 404 returned error can't find the container with id 8bbbcce9e1f4b38559aeca4b625686dfa611638475c6929685846a3d119b8486 Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.550181 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.570780 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.591468 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.611142 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.633615 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.650720 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.670200 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.691799 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.711766 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.716411 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k"] Nov 27 07:10:57 crc kubenswrapper[4706]: W1127 07:10:57.721423 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0558cfe_6d59_4897_9a10_595f42734ca6.slice/crio-4a5f70accc9925d8d4df8196684f11e0ad9e5f771785a755f7e5d8373cd24c96 WatchSource:0}: Error finding container 4a5f70accc9925d8d4df8196684f11e0ad9e5f771785a755f7e5d8373cd24c96: Status 404 returned error can't find the container with id 4a5f70accc9925d8d4df8196684f11e0ad9e5f771785a755f7e5d8373cd24c96 Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.735358 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.750449 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.771570 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.791128 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.811445 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.831013 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.851080 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.892108 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.910958 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.911184 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dd986c7f-e3ae-4d24-8815-c38d546168e7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c878d\" (UID: \"dd986c7f-e3ae-4d24-8815-c38d546168e7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.911347 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-key\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.911426 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba960b0a-5a50-4165-8572-0062d1ebff58-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rm5f8\" (UID: \"ba960b0a-5a50-4165-8572-0062d1ebff58\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.911456 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.911887 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-cabundle\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.911990 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.912095 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.912288 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94bada3b-c6c2-4e88-813c-becf573abbe2-cert\") pod \"ingress-canary-rns7p\" (UID: \"94bada3b-c6c2-4e88-813c-becf573abbe2\") " pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.912856 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-cabundle\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.913074 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.914949 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.917132 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-signing-key\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.917626 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94bada3b-c6c2-4e88-813c-becf573abbe2-cert\") pod \"ingress-canary-rns7p\" (UID: \"94bada3b-c6c2-4e88-813c-becf573abbe2\") " pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.919243 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dd986c7f-e3ae-4d24-8815-c38d546168e7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c878d\" (UID: \"dd986c7f-e3ae-4d24-8815-c38d546168e7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.919278 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba960b0a-5a50-4165-8572-0062d1ebff58-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rm5f8\" (UID: \"ba960b0a-5a50-4165-8572-0062d1ebff58\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.919333 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.930597 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.952395 4706 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.970960 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 07:10:57 crc kubenswrapper[4706]: I1127 07:10:57.991256 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.012408 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.030747 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.050397 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.086810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4xnk\" (UniqueName: \"kubernetes.io/projected/094c9ba3-fb86-48e5-93e0-f5f1f55cc12e-kube-api-access-f4xnk\") pod \"downloads-7954f5f757-dh98x\" (UID: \"094c9ba3-fb86-48e5-93e0-f5f1f55cc12e\") " pod="openshift-console/downloads-7954f5f757-dh98x" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.110258 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwj8d\" (UniqueName: \"kubernetes.io/projected/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-kube-api-access-xwj8d\") pod \"controller-manager-879f6c89f-lbldb\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.126064 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41f449f3-95c5-4193-81a2-ade651d73501-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.145147 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v48f\" (UniqueName: \"kubernetes.io/projected/1fbc1beb-c412-4ca8-ae88-8c73e17803b9-kube-api-access-5v48f\") pod \"cluster-samples-operator-665b6dd947-8ndx6\" (UID: \"1fbc1beb-c412-4ca8-ae88-8c73e17803b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.165098 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzp7g\" (UniqueName: \"kubernetes.io/projected/a6b3e5d2-bd7d-402d-af78-f685211e7f8e-kube-api-access-xzp7g\") pod \"dns-operator-744455d44c-n2f45\" (UID: \"a6b3e5d2-bd7d-402d-af78-f685211e7f8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.168400 4706 request.go:700] Waited for 1.906974747s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.190816 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6drrq\" (UniqueName: \"kubernetes.io/projected/036e6f98-ced3-4bb2-bc1f-f6005a15dd52-kube-api-access-6drrq\") pod \"console-operator-58897d9998-snqv9\" (UID: \"036e6f98-ced3-4bb2-bc1f-f6005a15dd52\") " pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.208779 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnltd\" (UniqueName: \"kubernetes.io/projected/0b4e82db-e890-4dc6-a6f0-441d78878459-kube-api-access-bnltd\") pod \"machine-api-operator-5694c8668f-r6h9j\" (UID: \"0b4e82db-e890-4dc6-a6f0-441d78878459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.221757 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.227362 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbkhn\" (UniqueName: \"kubernetes.io/projected/c9b6ede9-ef68-4b7b-a226-986c3c5fd6da-kube-api-access-rbkhn\") pod \"openshift-config-operator-7777fb866f-v8gbl\" (UID: \"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.232279 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.244628 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.254135 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w68w6\" (UniqueName: \"kubernetes.io/projected/ca2c33bd-a3b8-43da-86c2-d7221e5428c3-kube-api-access-w68w6\") pod \"machine-approver-56656f9798-kdm8j\" (UID: \"ca2c33bd-a3b8-43da-86c2-d7221e5428c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.260298 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.268476 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p27w\" (UniqueName: \"kubernetes.io/projected/41f449f3-95c5-4193-81a2-ade651d73501-kube-api-access-4p27w\") pod \"cluster-image-registry-operator-dc59b4c8b-q6249\" (UID: \"41f449f3-95c5-4193-81a2-ade651d73501\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.292217 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb4d0f97-a44e-4323-b52d-c060c71684ff-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pfq5q\" (UID: \"cb4d0f97-a44e-4323-b52d-c060c71684ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.297387 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dh98x" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.306153 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.313708 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.322779 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9h6d\" (UniqueName: \"kubernetes.io/projected/1d5dfadb-2cad-4379-bcdd-1578ac012cd6-kube-api-access-r9h6d\") pod \"etcd-operator-b45778765-mtl4k\" (UID: \"1d5dfadb-2cad-4379-bcdd-1578ac012cd6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.339160 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl2lb\" (UniqueName: \"kubernetes.io/projected/7f836197-8a96-4af0-b3bc-eb46a667b71c-kube-api-access-hl2lb\") pod \"apiserver-76f77b778f-nhj5d\" (UID: \"7f836197-8a96-4af0-b3bc-eb46a667b71c\") " pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.346692 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.355106 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cfzh\" (UniqueName: \"kubernetes.io/projected/d8e36cad-dbdd-458c-9862-7131d620ba34-kube-api-access-9cfzh\") pod \"console-f9d7485db-wt6br\" (UID: \"d8e36cad-dbdd-458c-9862-7131d620ba34\") " pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.361915 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.373343 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk9rz\" (UniqueName: \"kubernetes.io/projected/f7950dc1-2f65-4f07-9a5c-3f847effd16b-kube-api-access-hk9rz\") pod \"openshift-apiserver-operator-796bbdcf4f-2g4rq\" (UID: \"f7950dc1-2f65-4f07-9a5c-3f847effd16b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.389989 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvqph\" (UniqueName: \"kubernetes.io/projected/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-kube-api-access-dvqph\") pod \"oauth-openshift-558db77b4-fqglc\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.415723 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmhsn\" (UniqueName: \"kubernetes.io/projected/3939cbca-25db-434c-8527-ec3bc8e1aa85-kube-api-access-zmhsn\") pod \"openshift-controller-manager-operator-756b6f6bc6-2pkxz\" (UID: \"3939cbca-25db-434c-8527-ec3bc8e1aa85\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.434129 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w49c\" (UniqueName: \"kubernetes.io/projected/d91244b2-5d79-4540-a2ae-41ae33512da4-kube-api-access-4w49c\") pod \"machine-config-controller-84d6567774-xr4gg\" (UID: \"d91244b2-5d79-4540-a2ae-41ae33512da4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.440647 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.449840 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-957vs\" (UniqueName: \"kubernetes.io/projected/7d0da76e-b483-4a6a-a1b3-f4c64a29ddca-kube-api-access-957vs\") pod \"service-ca-9c57cc56f-np86m\" (UID: \"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca\") " pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.467121 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td7gw\" (UniqueName: \"kubernetes.io/projected/dd986c7f-e3ae-4d24-8815-c38d546168e7-kube-api-access-td7gw\") pod \"control-plane-machine-set-operator-78cbb6b69f-c878d\" (UID: \"dd986c7f-e3ae-4d24-8815-c38d546168e7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.487331 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj4n9\" (UniqueName: \"kubernetes.io/projected/ba960b0a-5a50-4165-8572-0062d1ebff58-kube-api-access-cj4n9\") pod \"multus-admission-controller-857f4d67dd-rm5f8\" (UID: \"ba960b0a-5a50-4165-8572-0062d1ebff58\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.497710 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.501590 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.503152 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" event={"ID":"f39b7c35-28bf-444f-a699-8607c48b6b62","Type":"ContainerStarted","Data":"88334bedc9510cf79f37b7077ec86a8d93655f74832d7113d46b486086438685"} Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.503232 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" event={"ID":"f39b7c35-28bf-444f-a699-8607c48b6b62","Type":"ContainerStarted","Data":"8bbbcce9e1f4b38559aeca4b625686dfa611638475c6929685846a3d119b8486"} Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.510063 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.510425 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3f56f0e-d18b-49a3-902a-3f97c03a75b9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hplkm\" (UID: \"f3f56f0e-d18b-49a3-902a-3f97c03a75b9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.510678 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" event={"ID":"a0558cfe-6d59-4897-9a10-595f42734ca6","Type":"ContainerStarted","Data":"0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f"} Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.510735 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" event={"ID":"a0558cfe-6d59-4897-9a10-595f42734ca6","Type":"ContainerStarted","Data":"4a5f70accc9925d8d4df8196684f11e0ad9e5f771785a755f7e5d8373cd24c96"} Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.511693 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.512893 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.521554 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.534559 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2bg9\" (UniqueName: \"kubernetes.io/projected/a709c580-4d9f-457d-a6cf-67a2831f4dcd-kube-api-access-h2bg9\") pod \"collect-profiles-29403780-gsqzn\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.541261 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-np86m" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.555269 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb7b9\" (UniqueName: \"kubernetes.io/projected/75bd3901-cc36-42ee-a1b5-78fd448ced38-kube-api-access-kb7b9\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.571070 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2snhf\" (UniqueName: \"kubernetes.io/projected/a4dcfcf5-d873-4adb-868a-58e1630b816a-kube-api-access-2snhf\") pod \"package-server-manager-789f6589d5-gdp66\" (UID: \"a4dcfcf5-d873-4adb-868a-58e1630b816a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.588953 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57pcz\" (UniqueName: \"kubernetes.io/projected/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-kube-api-access-57pcz\") pod \"marketplace-operator-79b997595-jgpht\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.598816 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.609053 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77b22\" (UniqueName: \"kubernetes.io/projected/94bada3b-c6c2-4e88-813c-becf573abbe2-kube-api-access-77b22\") pod \"ingress-canary-rns7p\" (UID: \"94bada3b-c6c2-4e88-813c-becf573abbe2\") " pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.622072 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.624396 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlwvh\" (UniqueName: \"kubernetes.io/projected/df22a91c-9a68-4b08-972f-397c91bbc2d9-kube-api-access-hlwvh\") pod \"apiserver-7bbb656c7d-nc99l\" (UID: \"df22a91c-9a68-4b08-972f-397c91bbc2d9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.631449 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.647921 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99cqs\" (UniqueName: \"kubernetes.io/projected/417efb56-71f3-4d60-bfb4-c1b11448f3be-kube-api-access-99cqs\") pod \"kube-storage-version-migrator-operator-b67b599dd-75k8h\" (UID: \"417efb56-71f3-4d60-bfb4-c1b11448f3be\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.655914 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.667568 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqxn9\" (UniqueName: \"kubernetes.io/projected/6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407-kube-api-access-qqxn9\") pod \"machine-config-operator-74547568cd-mscfh\" (UID: \"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.687277 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn7kx\" (UniqueName: \"kubernetes.io/projected/04db01d1-964d-49eb-9b17-4f81da5b31f9-kube-api-access-bn7kx\") pod \"olm-operator-6b444d44fb-pl2md\" (UID: \"04db01d1-964d-49eb-9b17-4f81da5b31f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.704905 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plh8h\" (UniqueName: \"kubernetes.io/projected/3aaba837-66f8-4b9e-a3d8-883406929734-kube-api-access-plh8h\") pod \"packageserver-d55dfcdfc-974lt\" (UID: \"3aaba837-66f8-4b9e-a3d8-883406929734\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.729080 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.732522 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6"] Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.735791 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75bd3901-cc36-42ee-a1b5-78fd448ced38-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7tfkr\" (UID: \"75bd3901-cc36-42ee-a1b5-78fd448ced38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.747285 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.749841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d8j2\" (UniqueName: \"kubernetes.io/projected/a9235278-ebf9-42f9-bfce-945f5453f26d-kube-api-access-8d8j2\") pod \"catalog-operator-68c6474976-95tgv\" (UID: \"a9235278-ebf9-42f9-bfce-945f5453f26d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.762790 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.764175 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.764467 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-snqv9"] Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.765428 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lbldb"] Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.767546 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.770048 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtcv7\" (UniqueName: \"kubernetes.io/projected/c7b9d8c1-2a4b-44f3-8d20-649b1c893018-kube-api-access-gtcv7\") pod \"migrator-59844c95c7-bmvl6\" (UID: \"c7b9d8c1-2a4b-44f3-8d20-649b1c893018\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.775612 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.798117 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.810493 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.825242 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.825369 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.825510 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-n2f45"] Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836527 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-certificates\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836571 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862955e9-b47d-4412-bac8-b087220b5228-config\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836619 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-tls\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836639 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g775f\" (UniqueName: \"kubernetes.io/projected/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-kube-api-access-g775f\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836673 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-trusted-ca\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836701 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-bound-sa-token\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836721 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-metrics-certs\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836773 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn8hc\" (UniqueName: \"kubernetes.io/projected/862955e9-b47d-4412-bac8-b087220b5228-kube-api-access-jn8hc\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836824 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-stats-auth\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836843 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54277327-8da6-4403-83e6-088e972bbe84-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836860 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862955e9-b47d-4412-bac8-b087220b5228-serving-cert\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836879 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54277327-8da6-4403-83e6-088e972bbe84-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836919 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bdec2b7f-560f-44e9-b665-2ff892495d19-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.836995 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54277327-8da6-4403-83e6-088e972bbe84-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.837065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bdec2b7f-560f-44e9-b665-2ff892495d19-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.837101 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.837120 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6fdj\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-kube-api-access-r6fdj\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.837135 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-default-certificate\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.837151 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-service-ca-bundle\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: E1127 07:10:58.840455 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.340440001 +0000 UTC m=+143.230030811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.846543 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.846978 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249"] Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.860968 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rns7p" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.937697 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:58 crc kubenswrapper[4706]: E1127 07:10:58.937891 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.437860525 +0000 UTC m=+143.327451335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.938062 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-csi-data-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.938367 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-tls\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.938705 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g775f\" (UniqueName: \"kubernetes.io/projected/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-kube-api-access-g775f\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.938928 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-trusted-ca\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.939023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-socket-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.939098 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-bound-sa-token\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.939174 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-metrics-certs\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.940054 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-trusted-ca\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.940512 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-registration-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.940728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn8hc\" (UniqueName: \"kubernetes.io/projected/862955e9-b47d-4412-bac8-b087220b5228-kube-api-access-jn8hc\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.940813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnk4v\" (UniqueName: \"kubernetes.io/projected/15940bb1-0f78-47e7-835c-54b72f82b2f6-kube-api-access-mnk4v\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.940917 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-mountpoint-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.940989 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcx4s\" (UniqueName: \"kubernetes.io/projected/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-kube-api-access-dcx4s\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.941097 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38a100a-2d88-428f-babd-d855eb0d65a7-config-volume\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.941167 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a38a100a-2d88-428f-babd-d855eb0d65a7-metrics-tls\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.943253 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54277327-8da6-4403-83e6-088e972bbe84-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.943362 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-stats-auth\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.943468 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgg8c\" (UniqueName: \"kubernetes.io/projected/a38a100a-2d88-428f-babd-d855eb0d65a7-kube-api-access-sgg8c\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.943578 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862955e9-b47d-4412-bac8-b087220b5228-serving-cert\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.943666 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-metrics-certs\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.943799 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54277327-8da6-4403-83e6-088e972bbe84-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.942265 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-tls\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.944018 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bdec2b7f-560f-44e9-b665-2ff892495d19-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.944140 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-plugins-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.944468 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/15940bb1-0f78-47e7-835c-54b72f82b2f6-certs\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.944624 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54277327-8da6-4403-83e6-088e972bbe84-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.944887 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bdec2b7f-560f-44e9-b665-2ff892495d19-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.945034 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6fdj\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-kube-api-access-r6fdj\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.945298 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-default-certificate\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.945378 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-service-ca-bundle\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.945453 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.945718 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-certificates\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.945832 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/15940bb1-0f78-47e7-835c-54b72f82b2f6-node-bootstrap-token\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.945944 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862955e9-b47d-4412-bac8-b087220b5228-config\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.947833 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bdec2b7f-560f-44e9-b665-2ff892495d19-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: E1127 07:10:58.949097 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.449082112 +0000 UTC m=+143.338672922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.950964 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-service-ca-bundle\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.953164 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/862955e9-b47d-4412-bac8-b087220b5228-serving-cert\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.953437 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862955e9-b47d-4412-bac8-b087220b5228-config\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.953922 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54277327-8da6-4403-83e6-088e972bbe84-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.954449 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-certificates\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.958079 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-stats-auth\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.963293 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bdec2b7f-560f-44e9-b665-2ff892495d19-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.963358 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54277327-8da6-4403-83e6-088e972bbe84-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.975459 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-default-certificate\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:58 crc kubenswrapper[4706]: I1127 07:10:58.988206 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g775f\" (UniqueName: \"kubernetes.io/projected/776f96c1-efaa-4cdd-9ef4-0533af4f2ca0-kube-api-access-g775f\") pod \"router-default-5444994796-8bwv8\" (UID: \"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0\") " pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.004579 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-bound-sa-token\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.009636 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.022987 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.034414 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn8hc\" (UniqueName: \"kubernetes.io/projected/862955e9-b47d-4412-bac8-b087220b5228-kube-api-access-jn8hc\") pod \"service-ca-operator-777779d784-tqzn8\" (UID: \"862955e9-b47d-4412-bac8-b087220b5228\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.039620 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.047400 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.048011 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54277327-8da6-4403-83e6-088e972bbe84-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mg8pf\" (UID: \"54277327-8da6-4403-83e6-088e972bbe84\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.048285 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.54826822 +0000 UTC m=+143.437859030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.048765 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-csi-data-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.048874 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-socket-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.048949 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-registration-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.049026 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnk4v\" (UniqueName: \"kubernetes.io/projected/15940bb1-0f78-47e7-835c-54b72f82b2f6-kube-api-access-mnk4v\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.049117 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcx4s\" (UniqueName: \"kubernetes.io/projected/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-kube-api-access-dcx4s\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.049208 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-mountpoint-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.049305 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a38a100a-2d88-428f-babd-d855eb0d65a7-metrics-tls\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.049375 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38a100a-2d88-428f-babd-d855eb0d65a7-config-volume\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.049446 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgg8c\" (UniqueName: \"kubernetes.io/projected/a38a100a-2d88-428f-babd-d855eb0d65a7-kube-api-access-sgg8c\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.049545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-plugins-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.050239 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-plugins-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.050415 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-csi-data-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.050455 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-socket-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.050482 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-registration-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.050644 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-mountpoint-dir\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.051254 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38a100a-2d88-428f-babd-d855eb0d65a7-config-volume\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.053021 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/15940bb1-0f78-47e7-835c-54b72f82b2f6-certs\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.053378 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.053722 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.553711697 +0000 UTC m=+143.443302507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.054453 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/15940bb1-0f78-47e7-835c-54b72f82b2f6-node-bootstrap-token\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.056264 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a38a100a-2d88-428f-babd-d855eb0d65a7-metrics-tls\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.064952 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6fdj\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-kube-api-access-r6fdj\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.065333 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/15940bb1-0f78-47e7-835c-54b72f82b2f6-certs\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.066830 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/15940bb1-0f78-47e7-835c-54b72f82b2f6-node-bootstrap-token\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.111258 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcx4s\" (UniqueName: \"kubernetes.io/projected/a56eee0b-d1b8-47e0-887f-4fe2d2ed47df-kube-api-access-dcx4s\") pod \"csi-hostpathplugin-8jrds\" (UID: \"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df\") " pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.130758 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.144190 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnk4v\" (UniqueName: \"kubernetes.io/projected/15940bb1-0f78-47e7-835c-54b72f82b2f6-kube-api-access-mnk4v\") pod \"machine-config-server-wcx4x\" (UID: \"15940bb1-0f78-47e7-835c-54b72f82b2f6\") " pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.150101 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mtl4k"] Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.155402 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgg8c\" (UniqueName: \"kubernetes.io/projected/a38a100a-2d88-428f-babd-d855eb0d65a7-kube-api-access-sgg8c\") pod \"dns-default-v9mq5\" (UID: \"a38a100a-2d88-428f-babd-d855eb0d65a7\") " pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.156398 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dh98x"] Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.156955 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.157087 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.657071967 +0000 UTC m=+143.546662777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.157447 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.157760 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.657752631 +0000 UTC m=+143.547343431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.162535 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r6h9j"] Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.167458 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v9mq5" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.191992 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.199692 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wcx4x" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.260023 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.260364 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.760346394 +0000 UTC m=+143.649937204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.335739 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.369041 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.369469 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.869453492 +0000 UTC m=+143.759044312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: W1127 07:10:59.373682 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod094c9ba3_fb86_48e5_93e0_f5f1f55cc12e.slice/crio-fc428fd2ff757dc1ffd5523449ffa50fa70cc35c3f0348a3a1f15e225ca3dec3 WatchSource:0}: Error finding container fc428fd2ff757dc1ffd5523449ffa50fa70cc35c3f0348a3a1f15e225ca3dec3: Status 404 returned error can't find the container with id fc428fd2ff757dc1ffd5523449ffa50fa70cc35c3f0348a3a1f15e225ca3dec3 Nov 27 07:10:59 crc kubenswrapper[4706]: W1127 07:10:59.379816 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15940bb1_0f78_47e7_835c_54b72f82b2f6.slice/crio-ff82e5ab09caa6f82c59685aeee104cd9e803febb19a56e9bfd516042bab047d WatchSource:0}: Error finding container ff82e5ab09caa6f82c59685aeee104cd9e803febb19a56e9bfd516042bab047d: Status 404 returned error can't find the container with id ff82e5ab09caa6f82c59685aeee104cd9e803febb19a56e9bfd516042bab047d Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.470851 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.471095 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.971064317 +0000 UTC m=+143.860655127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.471215 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.471547 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:10:59.971533624 +0000 UTC m=+143.861124434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.524039 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wcx4x" event={"ID":"15940bb1-0f78-47e7-835c-54b72f82b2f6","Type":"ContainerStarted","Data":"ff82e5ab09caa6f82c59685aeee104cd9e803febb19a56e9bfd516042bab047d"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.526119 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" event={"ID":"1d5dfadb-2cad-4379-bcdd-1578ac012cd6","Type":"ContainerStarted","Data":"d72a46d4f9c6b7b8e834b1b059038e0df06e5d8ade8ae2b4c56be1f62081eab4"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.528326 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8bwv8" event={"ID":"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0","Type":"ContainerStarted","Data":"bce45539f87c10cff5774c59ff8fe23f189fffea3e4f62764e402b9bc71f2fe5"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.532860 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" event={"ID":"50c0d8f5-2115-4b0c-957d-9e687cdb1d57","Type":"ContainerStarted","Data":"fa6704ca2eebb36a1ec03614f4f92d42f69a61126a66bc3cec80d4d3e52342ea"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.534744 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dh98x" event={"ID":"094c9ba3-fb86-48e5-93e0-f5f1f55cc12e","Type":"ContainerStarted","Data":"fc428fd2ff757dc1ffd5523449ffa50fa70cc35c3f0348a3a1f15e225ca3dec3"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.538563 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" event={"ID":"1fbc1beb-c412-4ca8-ae88-8c73e17803b9","Type":"ContainerStarted","Data":"f8b628a97a98160cd9558ed526b5bc8849132d9d1e32eeb6ff6c9537d83e864e"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.539925 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" event={"ID":"a6b3e5d2-bd7d-402d-af78-f685211e7f8e","Type":"ContainerStarted","Data":"9de7ca704f63cc366aec42902412d6391914adc6f1c45dc6c5208d8e6263857c"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.541779 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" event={"ID":"ca2c33bd-a3b8-43da-86c2-d7221e5428c3","Type":"ContainerStarted","Data":"20954d2da1557557d4d1e1cad27976446765413cf686adc70f91d1356a031247"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.541806 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" event={"ID":"ca2c33bd-a3b8-43da-86c2-d7221e5428c3","Type":"ContainerStarted","Data":"099034f38d6508cd4f34ef21495c148d8b969b8463681c88e236b5ce7041afed"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.544029 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" event={"ID":"0b4e82db-e890-4dc6-a6f0-441d78878459","Type":"ContainerStarted","Data":"bb7580ce3d5791360abae8f02ae9f9ddaf9ba8c263a1fbe26b94e51bedca1f9c"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.544944 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-snqv9" event={"ID":"036e6f98-ced3-4bb2-bc1f-f6005a15dd52","Type":"ContainerStarted","Data":"19c08309d0b9e8f4d7aa62acd7c0afad3ce6b659701f28a8c338c822f65bab8f"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.546764 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" event={"ID":"41f449f3-95c5-4193-81a2-ade651d73501","Type":"ContainerStarted","Data":"68e2bac6abcd29bdb9fe0e85d90a538efd6e36aeb75e9ac2523db80e4f8bf8a4"} Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.571969 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.572498 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.072479626 +0000 UTC m=+143.962070436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.652327 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q"] Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.674553 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.677031 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.177014298 +0000 UTC m=+144.066605108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.711259 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.777015 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.777670 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.277649919 +0000 UTC m=+144.167240729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.816802 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq"] Nov 27 07:10:59 crc kubenswrapper[4706]: W1127 07:10:59.876700 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7950dc1_2f65_4f07_9a5c_3f847effd16b.slice/crio-de36137f7e02862f491cc34493fba72c9efd2d55996b4242e7b04f2565de93bd WatchSource:0}: Error finding container de36137f7e02862f491cc34493fba72c9efd2d55996b4242e7b04f2565de93bd: Status 404 returned error can't find the container with id de36137f7e02862f491cc34493fba72c9efd2d55996b4242e7b04f2565de93bd Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.878955 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.879347 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.379333898 +0000 UTC m=+144.268924708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:10:59 crc kubenswrapper[4706]: I1127 07:10:59.981134 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:10:59 crc kubenswrapper[4706]: E1127 07:10:59.981651 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.481624888 +0000 UTC m=+144.371215698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.082731 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.083613 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.583585447 +0000 UTC m=+144.473176257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.150041 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.160102 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fqglc"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.179977 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.192851 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.193285 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.693265995 +0000 UTC m=+144.582856805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.194825 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.223008 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.225284 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rm5f8"] Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.232527 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04db01d1_964d_49eb_9b17_4f81da5b31f9.slice/crio-a6e25acfe799f708e44aa9760f572a92b7ef4c61c1f13152b8aff21ad518353b WatchSource:0}: Error finding container a6e25acfe799f708e44aa9760f572a92b7ef4c61c1f13152b8aff21ad518353b: Status 404 returned error can't find the container with id a6e25acfe799f708e44aa9760f572a92b7ef4c61c1f13152b8aff21ad518353b Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.235131 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ghm8f" podStartSLOduration=123.235113983 podStartE2EDuration="2m3.235113983s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.217540386 +0000 UTC m=+144.107131196" watchObservedRunningTime="2025-11-27 07:11:00.235113983 +0000 UTC m=+144.124704793" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.236353 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-np86m"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.252764 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.265913 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.275323 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-nhj5d"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.286882 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wt6br"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.288088 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh"] Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.288468 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba960b0a_5a50_4165_8572_0062d1ebff58.slice/crio-2e696178b76f940bc8e6fe4ed1fe1eb6b95556bcdca51a9171194c2b6bf57566 WatchSource:0}: Error finding container 2e696178b76f940bc8e6fe4ed1fe1eb6b95556bcdca51a9171194c2b6bf57566: Status 404 returned error can't find the container with id 2e696178b76f940bc8e6fe4ed1fe1eb6b95556bcdca51a9171194c2b6bf57566 Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.294780 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.295282 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.795268626 +0000 UTC m=+144.684859436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.302771 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.304548 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.308230 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.315461 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt"] Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.379988 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f836197_8a96_4af0_b3bc_eb46a667b71c.slice/crio-51078ea876f508b2e472c52986b90c3cd3bf607fd267473451d988f6fefb6def WatchSource:0}: Error finding container 51078ea876f508b2e472c52986b90c3cd3bf607fd267473451d988f6fefb6def: Status 404 returned error can't find the container with id 51078ea876f508b2e472c52986b90c3cd3bf607fd267473451d988f6fefb6def Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.396931 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.398391 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:00.898350835 +0000 UTC m=+144.787941665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.415793 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3aaba837_66f8_4b9e_a3d8_883406929734.slice/crio-059d42838ca20814a4c6b200a419fad5f76e9516bd8630410290d41368389072 WatchSource:0}: Error finding container 059d42838ca20814a4c6b200a419fad5f76e9516bd8630410290d41368389072: Status 404 returned error can't find the container with id 059d42838ca20814a4c6b200a419fad5f76e9516bd8630410290d41368389072 Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.441017 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v9mq5"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.445704 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf"] Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.473040 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda38a100a_2d88_428f_babd_d855eb0d65a7.slice/crio-b82cf9e616e4e91ee517a8b2b1abfcd21531fcd6257c805652a2a027a0327ac4 WatchSource:0}: Error finding container b82cf9e616e4e91ee517a8b2b1abfcd21531fcd6257c805652a2a027a0327ac4: Status 404 returned error can't find the container with id b82cf9e616e4e91ee517a8b2b1abfcd21531fcd6257c805652a2a027a0327ac4 Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.474097 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rns7p"] Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.474101 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54277327_8da6_4403_83e6_088e972bbe84.slice/crio-4898f710331394525f7a1ce1775474dd99399b9fb4dd0442fa9dc56f9a88e298 WatchSource:0}: Error finding container 4898f710331394525f7a1ce1775474dd99399b9fb4dd0442fa9dc56f9a88e298: Status 404 returned error can't find the container with id 4898f710331394525f7a1ce1775474dd99399b9fb4dd0442fa9dc56f9a88e298 Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.486728 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jgpht"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.491834 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.500641 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.501141 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.001125323 +0000 UTC m=+144.890716143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.523938 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6"] Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.549512 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75bd3901_cc36_42ee_a1b5_78fd448ced38.slice/crio-fc5befc0c3a0ad7cd97ab683e819e35a6ad17ef2e2c1c5f7ffae70490f95466c WatchSource:0}: Error finding container fc5befc0c3a0ad7cd97ab683e819e35a6ad17ef2e2c1c5f7ffae70490f95466c: Status 404 returned error can't find the container with id fc5befc0c3a0ad7cd97ab683e819e35a6ad17ef2e2c1c5f7ffae70490f95466c Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.553759 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.563032 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8jrds"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.601794 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.602230 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.102192209 +0000 UTC m=+144.991783019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.602693 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8"] Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.608534 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7b9d8c1_2a4b_44f3_8d20_649b1c893018.slice/crio-0379a442a8503e99951454e958e4985bbb6bbbb621dea7195fa3cec5566a8ce6 WatchSource:0}: Error finding container 0379a442a8503e99951454e958e4985bbb6bbbb621dea7195fa3cec5566a8ce6: Status 404 returned error can't find the container with id 0379a442a8503e99951454e958e4985bbb6bbbb621dea7195fa3cec5566a8ce6 Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.610639 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" event={"ID":"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407","Type":"ContainerStarted","Data":"e12152b10701f87089f88a7b0cd69e7b33ef983dcb9028f7a3ba387b300bf5f3"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.612814 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rns7p" event={"ID":"94bada3b-c6c2-4e88-813c-becf573abbe2","Type":"ContainerStarted","Data":"3b88f7bed20e5a02206904408da37a4b49cb6eca454c485b6e5b93ab45801d2e"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.619617 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8bwv8" event={"ID":"776f96c1-efaa-4cdd-9ef4-0533af4f2ca0","Type":"ContainerStarted","Data":"ca8604366d06bd839f49fb3633d0ba6c7a4c548b6179d80fb14c7a7078d1dfa3"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.624394 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66"] Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.633979 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wt6br" event={"ID":"d8e36cad-dbdd-458c-9862-7131d620ba34","Type":"ContainerStarted","Data":"52679674628df0935c79e32065455f38e6003ea92826c1c0d8171c7f08480705"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.654793 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-snqv9" event={"ID":"036e6f98-ced3-4bb2-bc1f-f6005a15dd52","Type":"ContainerStarted","Data":"851b8c56482e8a81a2be9979bbf043bd5a40bf506cb323b1376996942a7ecd03"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.655394 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.658233 4706 patch_prober.go:28] interesting pod/console-operator-58897d9998-snqv9 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.658273 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-snqv9" podUID="036e6f98-ced3-4bb2-bc1f-f6005a15dd52" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.658307 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" event={"ID":"e1ad5e9b-55f0-441a-84f7-d0f30410abe8","Type":"ContainerStarted","Data":"a4298ebea9a89eeedc168ff272328fb2566432a4c45cb02c34f4937152297f3c"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.662900 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" event={"ID":"ba960b0a-5a50-4165-8572-0062d1ebff58","Type":"ContainerStarted","Data":"2e696178b76f940bc8e6fe4ed1fe1eb6b95556bcdca51a9171194c2b6bf57566"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.664303 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" podStartSLOduration=122.664289392 podStartE2EDuration="2m2.664289392s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.637269162 +0000 UTC m=+144.526859982" watchObservedRunningTime="2025-11-27 07:11:00.664289392 +0000 UTC m=+144.553880202" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.681420 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" event={"ID":"e22ae6d0-dbdf-4b69-84f0-0643266a7dff","Type":"ContainerStarted","Data":"e95a9df2c02b5eadf2fd9615aac85af28a55f9771cb7fe3ffbdcfd8198b1ca70"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.683093 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" event={"ID":"f3f56f0e-d18b-49a3-902a-3f97c03a75b9","Type":"ContainerStarted","Data":"665c17295cd461e1d521fd4b04044ac69da923c351a9fcc56c686bcbfdc032aa"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.685641 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" event={"ID":"0b4e82db-e890-4dc6-a6f0-441d78878459","Type":"ContainerStarted","Data":"3c4818ffba3eb52a37fd08c2f8a9b39d767127acffd5ad2ca02e17b1b0ce6f5d"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.685663 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" event={"ID":"0b4e82db-e890-4dc6-a6f0-441d78878459","Type":"ContainerStarted","Data":"fec3f3142b2d7496dcf411f770ec8f6cb725f767b29e98abc350c3f3c2cddf2a"} Nov 27 07:11:00 crc kubenswrapper[4706]: W1127 07:11:00.687819 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4dcfcf5_d873_4adb_868a_58e1630b816a.slice/crio-05f1904aa002db2f08cc6b051e08d411aeb84282aaf58543fd4d10f0d24ba021 WatchSource:0}: Error finding container 05f1904aa002db2f08cc6b051e08d411aeb84282aaf58543fd4d10f0d24ba021: Status 404 returned error can't find the container with id 05f1904aa002db2f08cc6b051e08d411aeb84282aaf58543fd4d10f0d24ba021 Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.689418 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-np86m" event={"ID":"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca","Type":"ContainerStarted","Data":"5727bb57501fb274d5629303788033d50d477a64360a93fcd8dfd96673890eb4"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.696798 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wcx4x" event={"ID":"15940bb1-0f78-47e7-835c-54b72f82b2f6","Type":"ContainerStarted","Data":"a4ac94e3dda3b774cb1d9801899200445bb63b9cf4aa8f077556b5371889c048"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.698728 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" event={"ID":"a9235278-ebf9-42f9-bfce-945f5453f26d","Type":"ContainerStarted","Data":"c4c5e6df72cbdf029327d891b6dfa4e7fb6ce78da610c48c25b8b9afdb2ee4ac"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.703694 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.706648 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.206634278 +0000 UTC m=+145.096225078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.712617 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" event={"ID":"50c0d8f5-2115-4b0c-957d-9e687cdb1d57","Type":"ContainerStarted","Data":"382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.713310 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.716593 4706 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lbldb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.716643 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" podUID="50c0d8f5-2115-4b0c-957d-9e687cdb1d57" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.716798 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" event={"ID":"ca2c33bd-a3b8-43da-86c2-d7221e5428c3","Type":"ContainerStarted","Data":"9c207654735b83491e7bf882e55fa020ca61851450baea0d3d8a94f021e43d1d"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.718150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" event={"ID":"3939cbca-25db-434c-8527-ec3bc8e1aa85","Type":"ContainerStarted","Data":"0338547ffe7de0ecfa9455e9f37945c2956389b77a446d2a828b2fb92b09c262"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.722043 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" event={"ID":"1fbc1beb-c412-4ca8-ae88-8c73e17803b9","Type":"ContainerStarted","Data":"f06b36989f9a9699247d1e20bcaa4f3443e91a63ec3204dd6e1090ac545c96dc"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.722079 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" event={"ID":"1fbc1beb-c412-4ca8-ae88-8c73e17803b9","Type":"ContainerStarted","Data":"92a1f23744a2081bc13e7a41462a82ae45ac2dc27ed7dee7284ba4ebc76eab85"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.729042 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" event={"ID":"dd986c7f-e3ae-4d24-8815-c38d546168e7","Type":"ContainerStarted","Data":"ceaf73433b4bd7c60a4338117eeb615f857ff917663877e8fa0ac9afeece9d60"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.740106 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" event={"ID":"d91244b2-5d79-4540-a2ae-41ae33512da4","Type":"ContainerStarted","Data":"eeac2ec42d39d6e3975c86384b140d86cc5ca493904c2b27d69aedc4dc4cfdf2"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.740149 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" event={"ID":"d91244b2-5d79-4540-a2ae-41ae33512da4","Type":"ContainerStarted","Data":"b582b4d917f54883fe8e75ef72bdb6bc3f9fb2cf2281d26a889a2c4834a08545"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.752610 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dh98x" event={"ID":"094c9ba3-fb86-48e5-93e0-f5f1f55cc12e","Type":"ContainerStarted","Data":"c1ca8d9a469798ed356723e544cb859153010e7e7f6c37951e8f1f0f82f7d27a"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.752831 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-dh98x" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.755907 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-dh98x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.755964 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dh98x" podUID="094c9ba3-fb86-48e5-93e0-f5f1f55cc12e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.756486 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" event={"ID":"a6b3e5d2-bd7d-402d-af78-f685211e7f8e","Type":"ContainerStarted","Data":"42e5dcd73a263b7e62f0e7b55da1c7300e4a84e86164f4231819455f9ba07275"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.756540 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" event={"ID":"a6b3e5d2-bd7d-402d-af78-f685211e7f8e","Type":"ContainerStarted","Data":"6bcffa95e2d817f2f76638aaf2ec21c80bc79fa75351265d5c432155e487c9f4"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.758523 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" event={"ID":"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da","Type":"ContainerStarted","Data":"6df6f63ab44d25a446fb3e9cd6a3febd481376067bb3405f6d320d0c2e0e70ed"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.780436 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-r6h9j" podStartSLOduration=122.780418355 podStartE2EDuration="2m2.780418355s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.779982619 +0000 UTC m=+144.669573429" watchObservedRunningTime="2025-11-27 07:11:00.780418355 +0000 UTC m=+144.670009165" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.804636 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.804880 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.304853541 +0000 UTC m=+145.194444351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.805263 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.806007 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.305998583 +0000 UTC m=+145.195589393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.820191 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" event={"ID":"7f836197-8a96-4af0-b3bc-eb46a667b71c","Type":"ContainerStarted","Data":"51078ea876f508b2e472c52986b90c3cd3bf607fd267473451d988f6fefb6def"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.820263 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v9mq5" event={"ID":"a38a100a-2d88-428f-babd-d855eb0d65a7","Type":"ContainerStarted","Data":"b82cf9e616e4e91ee517a8b2b1abfcd21531fcd6257c805652a2a027a0327ac4"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.820278 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" event={"ID":"417efb56-71f3-4d60-bfb4-c1b11448f3be","Type":"ContainerStarted","Data":"2101abb07de8319404ed1dcf82e0f1deaa8003b1be2423a2974410c54cd5b16e"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.820291 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" event={"ID":"3aaba837-66f8-4b9e-a3d8-883406929734","Type":"ContainerStarted","Data":"059d42838ca20814a4c6b200a419fad5f76e9516bd8630410290d41368389072"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.825363 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-8bwv8" podStartSLOduration=123.825347545 podStartE2EDuration="2m3.825347545s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.822136038 +0000 UTC m=+144.711726848" watchObservedRunningTime="2025-11-27 07:11:00.825347545 +0000 UTC m=+144.714938355" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.841549 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" event={"ID":"1d5dfadb-2cad-4379-bcdd-1578ac012cd6","Type":"ContainerStarted","Data":"d253c43c9a433fc2e1e5f1a57a6fd85f20f732cefbf9707453bd0a978f3ab002"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.859969 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8ndx6" podStartSLOduration=123.85995201 podStartE2EDuration="2m3.85995201s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.859133431 +0000 UTC m=+144.748724251" watchObservedRunningTime="2025-11-27 07:11:00.85995201 +0000 UTC m=+144.749542820" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.868387 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" event={"ID":"04db01d1-964d-49eb-9b17-4f81da5b31f9","Type":"ContainerStarted","Data":"a6e25acfe799f708e44aa9760f572a92b7ef4c61c1f13152b8aff21ad518353b"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.868622 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.874535 4706 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pl2md container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.874579 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" podUID="04db01d1-964d-49eb-9b17-4f81da5b31f9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.875503 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" event={"ID":"cb4d0f97-a44e-4323-b52d-c060c71684ff","Type":"ContainerStarted","Data":"774711a70822e05c080bf369dd78a1e3312833217fbc986aeb15ffd74a1d52b4"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.875572 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" event={"ID":"cb4d0f97-a44e-4323-b52d-c060c71684ff","Type":"ContainerStarted","Data":"61d646fb019db6916d2b9d631b167fd71b6cf5fc64cb8e44ed51224bdb85892a"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.879279 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" event={"ID":"54277327-8da6-4403-83e6-088e972bbe84","Type":"ContainerStarted","Data":"4898f710331394525f7a1ce1775474dd99399b9fb4dd0442fa9dc56f9a88e298"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.880584 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" event={"ID":"a709c580-4d9f-457d-a6cf-67a2831f4dcd","Type":"ContainerStarted","Data":"ef4fbf08c7d28422dd568b7a65b1586888423eb0271407db4a055074d3b1b97f"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.898515 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" event={"ID":"f7950dc1-2f65-4f07-9a5c-3f847effd16b","Type":"ContainerStarted","Data":"a4714e980417be85d468665426c88f167f85d27af96ad76ebeffcfc4c002f9c4"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.898561 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" event={"ID":"f7950dc1-2f65-4f07-9a5c-3f847effd16b","Type":"ContainerStarted","Data":"de36137f7e02862f491cc34493fba72c9efd2d55996b4242e7b04f2565de93bd"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.902129 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" event={"ID":"41f449f3-95c5-4193-81a2-ade651d73501","Type":"ContainerStarted","Data":"abee18b0ebb61baa08f299cc4f2bf515eecde392b42fd9becd55a38af147e32a"} Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.911668 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:00 crc kubenswrapper[4706]: E1127 07:11:00.912996 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.412973663 +0000 UTC m=+145.302564473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.947852 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" podStartSLOduration=123.947830798 podStartE2EDuration="2m3.947830798s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.941402264 +0000 UTC m=+144.830993064" watchObservedRunningTime="2025-11-27 07:11:00.947830798 +0000 UTC m=+144.837421608" Nov 27 07:11:00 crc kubenswrapper[4706]: I1127 07:11:00.956638 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-wcx4x" podStartSLOduration=4.956610596 podStartE2EDuration="4.956610596s" podCreationTimestamp="2025-11-27 07:10:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.911303013 +0000 UTC m=+144.800893823" watchObservedRunningTime="2025-11-27 07:11:00.956610596 +0000 UTC m=+144.846201406" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.015138 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.016204 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.516188837 +0000 UTC m=+145.405779647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.026713 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.028562 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-snqv9" podStartSLOduration=124.028551076 podStartE2EDuration="2m4.028551076s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.022449224 +0000 UTC m=+144.912040044" watchObservedRunningTime="2025-11-27 07:11:01.028551076 +0000 UTC m=+144.918141886" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.029195 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdm8j" podStartSLOduration=124.029190119 podStartE2EDuration="2m4.029190119s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:00.984780058 +0000 UTC m=+144.874370878" watchObservedRunningTime="2025-11-27 07:11:01.029190119 +0000 UTC m=+144.918780929" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.034568 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:01 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:01 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:01 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.034613 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.063203 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-n2f45" podStartSLOduration=124.063098269 podStartE2EDuration="2m4.063098269s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.061833833 +0000 UTC m=+144.951424643" watchObservedRunningTime="2025-11-27 07:11:01.063098269 +0000 UTC m=+144.952689079" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.108340 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-dh98x" podStartSLOduration=124.10832393 podStartE2EDuration="2m4.10832393s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.105958793 +0000 UTC m=+144.995549603" watchObservedRunningTime="2025-11-27 07:11:01.10832393 +0000 UTC m=+144.997914740" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.117774 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.118195 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.618153826 +0000 UTC m=+145.507744636 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.118562 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.118957 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.618949515 +0000 UTC m=+145.508540325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.183359 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pfq5q" podStartSLOduration=124.183337941 podStartE2EDuration="2m4.183337941s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.148738696 +0000 UTC m=+145.038329506" watchObservedRunningTime="2025-11-27 07:11:01.183337941 +0000 UTC m=+145.072928751" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.183715 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2g4rq" podStartSLOduration=124.183706305 podStartE2EDuration="2m4.183706305s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.179842744 +0000 UTC m=+145.069433554" watchObservedRunningTime="2025-11-27 07:11:01.183706305 +0000 UTC m=+145.073297115" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.220013 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.220407 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.720389045 +0000 UTC m=+145.609979855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.242960 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" podStartSLOduration=123.242936092 podStartE2EDuration="2m3.242936092s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.228977666 +0000 UTC m=+145.118568486" watchObservedRunningTime="2025-11-27 07:11:01.242936092 +0000 UTC m=+145.132526892" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.271352 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6249" podStartSLOduration=124.271331513 podStartE2EDuration="2m4.271331513s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.267298757 +0000 UTC m=+145.156889577" watchObservedRunningTime="2025-11-27 07:11:01.271331513 +0000 UTC m=+145.160922323" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.306565 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mtl4k" podStartSLOduration=124.306551171 podStartE2EDuration="2m4.306551171s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.305757962 +0000 UTC m=+145.195348782" watchObservedRunningTime="2025-11-27 07:11:01.306551171 +0000 UTC m=+145.196141981" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.322115 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.322501 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.822488978 +0000 UTC m=+145.712079788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.423898 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.424093 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.924044493 +0000 UTC m=+145.813635303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.424719 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.425279 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:01.925258276 +0000 UTC m=+145.814849086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.526408 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.526756 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.026722377 +0000 UTC m=+145.916313187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.527158 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.527781 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.027762745 +0000 UTC m=+145.917353555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.631267 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.631393 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.131375164 +0000 UTC m=+146.020965974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.631811 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.632385 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.132369289 +0000 UTC m=+146.021960099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.732801 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.734203 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.234182913 +0000 UTC m=+146.123773723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.734252 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.734541 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.234531516 +0000 UTC m=+146.124122326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.835203 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.835420 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.335363783 +0000 UTC m=+146.224954633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.835659 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.836023 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.336006747 +0000 UTC m=+146.225597567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.910107 4706 generic.go:334] "Generic (PLEG): container finished" podID="7f836197-8a96-4af0-b3bc-eb46a667b71c" containerID="a38205b7a4941f38c416dee196b0abced54f6545c5ba5a39663caa3606579b96" exitCode=0 Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.910369 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" event={"ID":"7f836197-8a96-4af0-b3bc-eb46a667b71c","Type":"ContainerDied","Data":"a38205b7a4941f38c416dee196b0abced54f6545c5ba5a39663caa3606579b96"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.912046 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" event={"ID":"a4dcfcf5-d873-4adb-868a-58e1630b816a","Type":"ContainerStarted","Data":"514158cdb204cac0e09362f1247cf5ea511c57182369bee75ba0f798b03a761d"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.912092 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" event={"ID":"a4dcfcf5-d873-4adb-868a-58e1630b816a","Type":"ContainerStarted","Data":"05f1904aa002db2f08cc6b051e08d411aeb84282aaf58543fd4d10f0d24ba021"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.915297 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" event={"ID":"3aaba837-66f8-4b9e-a3d8-883406929734","Type":"ContainerStarted","Data":"822491069d346f440bd8acca65d2ba35e2c479c976ed7401c2df602aafe68446"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.915710 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.917583 4706 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-974lt container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.917619 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" podUID="3aaba837-66f8-4b9e-a3d8-883406929734" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.925469 4706 generic.go:334] "Generic (PLEG): container finished" podID="df22a91c-9a68-4b08-972f-397c91bbc2d9" containerID="4f18c68dc0baf002633c2992dc0c23bd221dbe7658a972f18c300b9685b25edb" exitCode=0 Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.925925 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" event={"ID":"df22a91c-9a68-4b08-972f-397c91bbc2d9","Type":"ContainerDied","Data":"4f18c68dc0baf002633c2992dc0c23bd221dbe7658a972f18c300b9685b25edb"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.926007 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" event={"ID":"df22a91c-9a68-4b08-972f-397c91bbc2d9","Type":"ContainerStarted","Data":"11d4293e4d7db3682629b95904061d4f03fce017513c2a416c6147797786eb5e"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.933198 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" event={"ID":"3939cbca-25db-434c-8527-ec3bc8e1aa85","Type":"ContainerStarted","Data":"5acb8138e1973aa0ef5a40805cc003cfaf17fff885ecc7f87b3e8455812ef172"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.938618 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:01 crc kubenswrapper[4706]: E1127 07:11:01.939008 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.438990903 +0000 UTC m=+146.328581713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.955841 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wt6br" event={"ID":"d8e36cad-dbdd-458c-9862-7131d620ba34","Type":"ContainerStarted","Data":"3e18072d0d16053220a6cd838a96e83bd377ff4fd94853128541dde847cf8602"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.962827 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2pkxz" podStartSLOduration=124.962807756 podStartE2EDuration="2m4.962807756s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:01.960611376 +0000 UTC m=+145.850202186" watchObservedRunningTime="2025-11-27 07:11:01.962807756 +0000 UTC m=+145.852398566" Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.966448 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" event={"ID":"862955e9-b47d-4412-bac8-b087220b5228","Type":"ContainerStarted","Data":"ccffee98f7541c73b8f1887b3fc18db0f555c227e1c1baf607d5a731a9243090"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.966489 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" event={"ID":"862955e9-b47d-4412-bac8-b087220b5228","Type":"ContainerStarted","Data":"9bf573cd2a27a3a4f7d967b0c79102bf65d5b2378c77e1d8eb9ef41e62bda73f"} Nov 27 07:11:01 crc kubenswrapper[4706]: I1127 07:11:01.980949 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-np86m" event={"ID":"7d0da76e-b483-4a6a-a1b3-f4c64a29ddca","Type":"ContainerStarted","Data":"8bab086ef9c644759715914337e67e23cd01035d5ac6cb8c248bb1ef9b29bd26"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.034749 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" event={"ID":"f3f56f0e-d18b-49a3-902a-3f97c03a75b9","Type":"ContainerStarted","Data":"f7447ae76688075637a7ea635a2d42a774d26f60621754f85d16c78eb9f3cb56"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.034932 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:02 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:02 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:02 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.035863 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.040409 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.042087 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.542073942 +0000 UTC m=+146.431664752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.051801 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" event={"ID":"75bd3901-cc36-42ee-a1b5-78fd448ced38","Type":"ContainerStarted","Data":"7df33503e62fc56228cd4d8e8ec593dac483c049138514038370236897d15bf8"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.051843 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" event={"ID":"75bd3901-cc36-42ee-a1b5-78fd448ced38","Type":"ContainerStarted","Data":"fc5befc0c3a0ad7cd97ab683e819e35a6ad17ef2e2c1c5f7ffae70490f95466c"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.051927 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" podStartSLOduration=124.051911539 podStartE2EDuration="2m4.051911539s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.051342328 +0000 UTC m=+145.940933138" watchObservedRunningTime="2025-11-27 07:11:02.051911539 +0000 UTC m=+145.941502349" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.094607 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" event={"ID":"417efb56-71f3-4d60-bfb4-c1b11448f3be","Type":"ContainerStarted","Data":"c23977804fecd7f95c0f5affd0b192d5d0dce487453392bc2db5a9a71b0f4028"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.108197 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" event={"ID":"a9235278-ebf9-42f9-bfce-945f5453f26d","Type":"ContainerStarted","Data":"b162e1bb367443eaee119cb402963e9b1706975c52a76aebdc8c287fdf1be43a"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.109155 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.115329 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" event={"ID":"c7b9d8c1-2a4b-44f3-8d20-649b1c893018","Type":"ContainerStarted","Data":"871257ec71b65a7bd5b5e015e5e801745b8dd7676f6286d5647df31e57fb3b9a"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.115369 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" event={"ID":"c7b9d8c1-2a4b-44f3-8d20-649b1c893018","Type":"ContainerStarted","Data":"0379a442a8503e99951454e958e4985bbb6bbbb621dea7195fa3cec5566a8ce6"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.127907 4706 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-95tgv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.127966 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" podUID="a9235278-ebf9-42f9-bfce-945f5453f26d" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.128552 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hplkm" podStartSLOduration=125.128528198 podStartE2EDuration="2m5.128528198s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.093493567 +0000 UTC m=+145.983084377" watchObservedRunningTime="2025-11-27 07:11:02.128528198 +0000 UTC m=+146.018119018" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.133355 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wt6br" podStartSLOduration=125.133335272 podStartE2EDuration="2m5.133335272s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.129317487 +0000 UTC m=+146.018908297" watchObservedRunningTime="2025-11-27 07:11:02.133335272 +0000 UTC m=+146.022926082" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.134107 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" event={"ID":"54277327-8da6-4403-83e6-088e972bbe84","Type":"ContainerStarted","Data":"3359716f00df501064a62fd4c6923222955f2d19f86d3bbdd9df0fe58f8ab65f"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.138556 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" event={"ID":"04db01d1-964d-49eb-9b17-4f81da5b31f9","Type":"ContainerStarted","Data":"964b484c266149d8676a3131717879468584599b0f66581fb112614d82ab0897"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.139056 4706 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pl2md container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.139090 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" podUID="04db01d1-964d-49eb-9b17-4f81da5b31f9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.142157 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" event={"ID":"e22ae6d0-dbdf-4b69-84f0-0643266a7dff","Type":"ContainerStarted","Data":"11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.143090 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.143767 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.143859 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.643834993 +0000 UTC m=+146.533425863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.146359 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.149504 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.649488729 +0000 UTC m=+146.539079629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.150724 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqzn8" podStartSLOduration=124.150707472 podStartE2EDuration="2m4.150707472s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.149480488 +0000 UTC m=+146.039071298" watchObservedRunningTime="2025-11-27 07:11:02.150707472 +0000 UTC m=+146.040298282" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.157287 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" event={"ID":"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df","Type":"ContainerStarted","Data":"bed93671878533ee876954b1b11ab526a0861f2c067bcf810aa8c13cf0892706"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.171992 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" event={"ID":"a709c580-4d9f-457d-a6cf-67a2831f4dcd","Type":"ContainerStarted","Data":"c6165ffbdbb17cdd24d01e9499fad1d374c187c0cc66efc9cd2ba3649dbd76a6"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.173431 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jgpht container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.173501 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.186202 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-np86m" podStartSLOduration=124.1861842 podStartE2EDuration="2m4.1861842s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.183911327 +0000 UTC m=+146.073502137" watchObservedRunningTime="2025-11-27 07:11:02.1861842 +0000 UTC m=+146.075775010" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.233410 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" event={"ID":"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407","Type":"ContainerStarted","Data":"e896930bcf4e633f89869056b37f6e71c9ddd649472c8d88e8803824bc0420b3"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.233461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" event={"ID":"6dd9a6bc-1bc6-4cdb-9b3c-be548eefd407","Type":"ContainerStarted","Data":"ef63b14dc23a6059e12aafd2e55d757416bda614063161ddb9fed3f41a13da11"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.251051 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.252555 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" podStartSLOduration=124.252539497 podStartE2EDuration="2m4.252539497s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.250489262 +0000 UTC m=+146.140080072" watchObservedRunningTime="2025-11-27 07:11:02.252539497 +0000 UTC m=+146.142130297" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.252693 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" podStartSLOduration=124.252689432 podStartE2EDuration="2m4.252689432s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.227329512 +0000 UTC m=+146.116920412" watchObservedRunningTime="2025-11-27 07:11:02.252689432 +0000 UTC m=+146.142280242" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.252732 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.752702943 +0000 UTC m=+146.642293793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.271716 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v9mq5" event={"ID":"a38a100a-2d88-428f-babd-d855eb0d65a7","Type":"ContainerStarted","Data":"5b3834830b76da38414086d99940c71b4892204ca71ed0337565655ce55fbda5"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.271772 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v9mq5" event={"ID":"a38a100a-2d88-428f-babd-d855eb0d65a7","Type":"ContainerStarted","Data":"9c9b38dfb12ddd274465abd6828cb6430e91da567167ae7434694846769042d1"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.271868 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-v9mq5" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.280511 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" event={"ID":"e1ad5e9b-55f0-441a-84f7-d0f30410abe8","Type":"ContainerStarted","Data":"8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.281163 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.286532 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" event={"ID":"dd986c7f-e3ae-4d24-8815-c38d546168e7","Type":"ContainerStarted","Data":"a3f72309ba5c3ace081a1b0a6742940b9f037405bee118ddf574b57eefa4fae7"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.296919 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" podStartSLOduration=125.296903276 podStartE2EDuration="2m5.296903276s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.294327042 +0000 UTC m=+146.183917842" watchObservedRunningTime="2025-11-27 07:11:02.296903276 +0000 UTC m=+146.186494086" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.297476 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-75k8h" podStartSLOduration=125.297471217 podStartE2EDuration="2m5.297471217s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.278091364 +0000 UTC m=+146.167682174" watchObservedRunningTime="2025-11-27 07:11:02.297471217 +0000 UTC m=+146.187062017" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.311101 4706 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fqglc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.311163 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" podUID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.314736 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" event={"ID":"d91244b2-5d79-4540-a2ae-41ae33512da4","Type":"ContainerStarted","Data":"ab080ee0b08e13058f308dba7fe98cc5ac3b416015355725accea68249ec22ef"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.322846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" event={"ID":"ba960b0a-5a50-4165-8572-0062d1ebff58","Type":"ContainerStarted","Data":"f96eee15d4cba5b782eac5aa96c2894d6a4a6750e4f2e134553f69c1c87a036f"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.322885 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" event={"ID":"ba960b0a-5a50-4165-8572-0062d1ebff58","Type":"ContainerStarted","Data":"28322d2025f6f9dca4bbfec58894d71658d0e98acdbf233ec98f7d7f494620ee"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.329482 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mg8pf" podStartSLOduration=125.329469577 podStartE2EDuration="2m5.329469577s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.32678914 +0000 UTC m=+146.216379950" watchObservedRunningTime="2025-11-27 07:11:02.329469577 +0000 UTC m=+146.219060387" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.350638 4706 generic.go:334] "Generic (PLEG): container finished" podID="c9b6ede9-ef68-4b7b-a226-986c3c5fd6da" containerID="98522b628953a3d47319c952c481af70e83851a4d9951fb81e7f475b90900d4e" exitCode=0 Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.350744 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" event={"ID":"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da","Type":"ContainerDied","Data":"98522b628953a3d47319c952c481af70e83851a4d9951fb81e7f475b90900d4e"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.352701 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4gg" podStartSLOduration=124.352690779 podStartE2EDuration="2m4.352690779s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.352031665 +0000 UTC m=+146.241622475" watchObservedRunningTime="2025-11-27 07:11:02.352690779 +0000 UTC m=+146.242281589" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.359343 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.359641 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.859628992 +0000 UTC m=+146.749219802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.405182 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rns7p" event={"ID":"94bada3b-c6c2-4e88-813c-becf573abbe2","Type":"ContainerStarted","Data":"0170ec37fa3235a6c43c0e70884ace6c15085edce866295e85ff5fad37ddd785"} Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.405386 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-dh98x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.405438 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dh98x" podUID="094c9ba3-fb86-48e5-93e0-f5f1f55cc12e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.428628 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.447907 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-v9mq5" podStartSLOduration=7.447889143 podStartE2EDuration="7.447889143s" podCreationTimestamp="2025-11-27 07:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.441145198 +0000 UTC m=+146.330736008" watchObservedRunningTime="2025-11-27 07:11:02.447889143 +0000 UTC m=+146.337479953" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.460767 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.461945 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:02.961927752 +0000 UTC m=+146.851518562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.565148 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.565609 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.065592383 +0000 UTC m=+146.955183203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.586354 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" podStartSLOduration=125.586337215 podStartE2EDuration="2m5.586337215s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.585761034 +0000 UTC m=+146.475351864" watchObservedRunningTime="2025-11-27 07:11:02.586337215 +0000 UTC m=+146.475928035" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.666792 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.667198 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.167172088 +0000 UTC m=+147.056762908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.667624 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.667945 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.167933304 +0000 UTC m=+147.057524114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.672501 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rm5f8" podStartSLOduration=124.67248214 podStartE2EDuration="2m4.67248214s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.671136271 +0000 UTC m=+146.560727081" watchObservedRunningTime="2025-11-27 07:11:02.67248214 +0000 UTC m=+146.562072950" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.704301 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mscfh" podStartSLOduration=124.704270002 podStartE2EDuration="2m4.704270002s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.702107534 +0000 UTC m=+146.591698344" watchObservedRunningTime="2025-11-27 07:11:02.704270002 +0000 UTC m=+146.593860812" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.736879 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c878d" podStartSLOduration=124.736846615 podStartE2EDuration="2m4.736846615s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.73617424 +0000 UTC m=+146.625765050" watchObservedRunningTime="2025-11-27 07:11:02.736846615 +0000 UTC m=+146.626437425" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.744758 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-snqv9" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.769007 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.769483 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.269460828 +0000 UTC m=+147.159051638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.881313 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.881499 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.381473831 +0000 UTC m=+147.271064641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.914616 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rns7p" podStartSLOduration=7.914599602 podStartE2EDuration="7.914599602s" podCreationTimestamp="2025-11-27 07:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:02.913592156 +0000 UTC m=+146.803182966" watchObservedRunningTime="2025-11-27 07:11:02.914599602 +0000 UTC m=+146.804190412" Nov 27 07:11:02 crc kubenswrapper[4706]: I1127 07:11:02.986912 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:02 crc kubenswrapper[4706]: E1127 07:11:02.987298 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.487280119 +0000 UTC m=+147.376870929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.033693 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:03 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:03 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:03 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.034116 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.090505 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.090846 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.590829235 +0000 UTC m=+147.480420045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.191297 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.191863 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.691830199 +0000 UTC m=+147.581421009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.293396 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.293808 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.793790398 +0000 UTC m=+147.683381208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.394477 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.394628 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.894603565 +0000 UTC m=+147.784194395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.394791 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.395114 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.895104564 +0000 UTC m=+147.784695374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.411526 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" event={"ID":"df22a91c-9a68-4b08-972f-397c91bbc2d9","Type":"ContainerStarted","Data":"ec28c9da7a01aacc4b69558155be9e06f9c5ab2fa1c8b8d2498e90a21766fb16"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.414674 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" event={"ID":"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df","Type":"ContainerStarted","Data":"de304dca8604c77bccfb97321f46165b3bea38411203fade3ead7239bcf53369"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.417491 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" event={"ID":"c9b6ede9-ef68-4b7b-a226-986c3c5fd6da","Type":"ContainerStarted","Data":"6e647c97322bb1bae71d98d717695073c92e0e810031adce74a1e336a4e5fb14"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.418206 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.420145 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" event={"ID":"75bd3901-cc36-42ee-a1b5-78fd448ced38","Type":"ContainerStarted","Data":"582b66bb26438a75bbcbda6fbf813fbd49a30f47cfb5811728055911af49a545"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.423149 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" event={"ID":"c7b9d8c1-2a4b-44f3-8d20-649b1c893018","Type":"ContainerStarted","Data":"5eb374c1d4109750949d2f9141024b81dbe351cb9af491b8191fc99799fa394f"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.425746 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" event={"ID":"7f836197-8a96-4af0-b3bc-eb46a667b71c","Type":"ContainerStarted","Data":"e93904073373de585b77f813a35cec3e6977f639a8696e0ba19b6a1910d3e2f8"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.425774 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" event={"ID":"7f836197-8a96-4af0-b3bc-eb46a667b71c","Type":"ContainerStarted","Data":"ab7b238f09fc9e82a7ac9447e6dcfeeed5caca212ab256fd8eeea118fd0a33ba"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.427582 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" event={"ID":"a4dcfcf5-d873-4adb-868a-58e1630b816a","Type":"ContainerStarted","Data":"92b4be801e62b2e7339a58cdb145f09f3cb80d425539ca2e8196c6b725e9fea0"} Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.429505 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jgpht container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.429560 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.430396 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.469116 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-95tgv" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.488487 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pl2md" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.496346 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.496699 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:03.996661447 +0000 UTC m=+147.886252257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.497761 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.498503 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" podStartSLOduration=125.498484574 podStartE2EDuration="2m5.498484574s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:03.497047661 +0000 UTC m=+147.386638471" watchObservedRunningTime="2025-11-27 07:11:03.498484574 +0000 UTC m=+147.388075384" Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.503788 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.003767685 +0000 UTC m=+147.893358695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.550687 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" podStartSLOduration=125.550662416 podStartE2EDuration="2m5.550662416s" podCreationTimestamp="2025-11-27 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:03.544022236 +0000 UTC m=+147.433613046" watchObservedRunningTime="2025-11-27 07:11:03.550662416 +0000 UTC m=+147.440253226" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.578338 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" podStartSLOduration=126.578319329 podStartE2EDuration="2m6.578319329s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:03.576594277 +0000 UTC m=+147.466185087" watchObservedRunningTime="2025-11-27 07:11:03.578319329 +0000 UTC m=+147.467910139" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.599493 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.599781 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.599807 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.599887 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.099865221 +0000 UTC m=+147.989456031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.600598 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.601067 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.101056594 +0000 UTC m=+147.990647414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.601870 4706 patch_prober.go:28] interesting pod/apiserver-76f77b778f-nhj5d container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.12:8443/livez\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.601960 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" podUID="7f836197-8a96-4af0-b3bc-eb46a667b71c" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.12:8443/livez\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.638269 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bmvl6" podStartSLOduration=126.638238553 podStartE2EDuration="2m6.638238553s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:03.620893984 +0000 UTC m=+147.510484794" watchObservedRunningTime="2025-11-27 07:11:03.638238553 +0000 UTC m=+147.527829363" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.701865 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.702059 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.202038607 +0000 UTC m=+148.091629407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.702196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.702626 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.202615508 +0000 UTC m=+148.092206318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.765274 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.765716 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.773680 4706 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-nc99l container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.37:8443/livez\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.773732 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" podUID="df22a91c-9a68-4b08-972f-397c91bbc2d9" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.37:8443/livez\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.775046 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" podStartSLOduration=126.775033135 podStartE2EDuration="2m6.775033135s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:03.768329422 +0000 UTC m=+147.657920242" watchObservedRunningTime="2025-11-27 07:11:03.775033135 +0000 UTC m=+147.664623945" Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.806805 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.807635 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.307614177 +0000 UTC m=+148.197204987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.912405 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:03 crc kubenswrapper[4706]: E1127 07:11:03.912863 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.412847815 +0000 UTC m=+148.302438625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:03 crc kubenswrapper[4706]: I1127 07:11:03.915969 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7tfkr" podStartSLOduration=126.915939917 podStartE2EDuration="2m6.915939917s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:03.858178061 +0000 UTC m=+147.747768861" watchObservedRunningTime="2025-11-27 07:11:03.915939917 +0000 UTC m=+147.805530737" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.014238 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.014706 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.514642297 +0000 UTC m=+148.404233107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.014914 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.015336 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.515317651 +0000 UTC m=+148.404908461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.027268 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:04 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:04 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:04 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.027329 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.116151 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.116539 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.616501412 +0000 UTC m=+148.506092222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.116836 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.117144 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.617131765 +0000 UTC m=+148.506722575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.218172 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.218540 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.718510263 +0000 UTC m=+148.608101073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.218688 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.219027 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.71901483 +0000 UTC m=+148.608605640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.320232 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.320368 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.820337546 +0000 UTC m=+148.709928356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.320532 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.320825 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.820813254 +0000 UTC m=+148.710404064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.421785 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.422351 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:04.922333666 +0000 UTC m=+148.811924476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.429133 4706 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fqglc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": context deadline exceeded" start-of-body= Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.429161 4706 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-974lt container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.429214 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" podUID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": context deadline exceeded" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.429301 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" podUID="3aaba837-66f8-4b9e-a3d8-883406929734" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.436905 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" event={"ID":"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df","Type":"ContainerStarted","Data":"86cad323162ef9a21832d4f2dc2d4a43ad751dc8b7b9167a0ef34c0b82918a0e"} Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.438632 4706 generic.go:334] "Generic (PLEG): container finished" podID="a709c580-4d9f-457d-a6cf-67a2831f4dcd" containerID="c6165ffbdbb17cdd24d01e9499fad1d374c187c0cc66efc9cd2ba3649dbd76a6" exitCode=0 Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.439202 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" event={"ID":"a709c580-4d9f-457d-a6cf-67a2831f4dcd","Type":"ContainerDied","Data":"c6165ffbdbb17cdd24d01e9499fad1d374c187c0cc66efc9cd2ba3649dbd76a6"} Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.446017 4706 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jgpht container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.446061 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.523281 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.524691 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.024669228 +0000 UTC m=+148.914260108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.589125 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.624776 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.625009 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.625049 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.625612 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.125579939 +0000 UTC m=+149.015170759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.626809 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.632240 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.725915 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.726265 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.726301 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.727238 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.227203586 +0000 UTC m=+149.116794396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.734593 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.742945 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.810518 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.829783 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.833313 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.833934 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.333912767 +0000 UTC m=+149.223503587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.844370 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 07:11:04 crc kubenswrapper[4706]: I1127 07:11:04.936082 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:04 crc kubenswrapper[4706]: E1127 07:11:04.936394 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.436380793 +0000 UTC m=+149.325971593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.028678 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:05 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:05 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:05 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.028982 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.037084 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.037458 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.537441119 +0000 UTC m=+149.427031929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.076170 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tq6jq"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.077425 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.101159 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.107808 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tq6jq"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.138347 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.138730 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.638717223 +0000 UTC m=+149.528308033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.241991 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.242400 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-utilities\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.242502 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5d64\" (UniqueName: \"kubernetes.io/projected/43c8fce5-14bd-4d14-b488-50600b596cb5-kube-api-access-l5d64\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.242544 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-catalog-content\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.242688 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.742664004 +0000 UTC m=+149.632254814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.286350 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jj2k9"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.296304 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.305775 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.343867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5d64\" (UniqueName: \"kubernetes.io/projected/43c8fce5-14bd-4d14-b488-50600b596cb5-kube-api-access-l5d64\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.344172 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-catalog-content\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.344229 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-utilities\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.344252 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.344542 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.844530009 +0000 UTC m=+149.734120819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.345249 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-catalog-content\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.345474 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-utilities\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.361233 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj2k9"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.389529 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5d64\" (UniqueName: \"kubernetes.io/projected/43c8fce5-14bd-4d14-b488-50600b596cb5-kube-api-access-l5d64\") pod \"community-operators-tq6jq\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.440010 4706 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-974lt container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.440074 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" podUID="3aaba837-66f8-4b9e-a3d8-883406929734" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.444482 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.445426 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.445677 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-utilities\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.445743 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-catalog-content\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.445777 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wcqm\" (UniqueName: \"kubernetes.io/projected/a8ac4a84-ddce-41e8-9da0-efd60117aa67-kube-api-access-8wcqm\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.445872 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:05.945855925 +0000 UTC m=+149.835446735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.447154 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" event={"ID":"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df","Type":"ContainerStarted","Data":"73110081c1487dc84fe9ba80ee448f4afa20c179bfa1332244c699bd080ab4aa"} Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.483323 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w77l6"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.484294 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.539646 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w77l6"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.552278 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.552316 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wcqm\" (UniqueName: \"kubernetes.io/projected/a8ac4a84-ddce-41e8-9da0-efd60117aa67-kube-api-access-8wcqm\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.552437 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-utilities\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.552696 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-catalog-content\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.557188 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-utilities\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.557357 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-catalog-content\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.557755 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.057736573 +0000 UTC m=+149.947327383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.593425 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wcqm\" (UniqueName: \"kubernetes.io/projected/a8ac4a84-ddce-41e8-9da0-efd60117aa67-kube-api-access-8wcqm\") pod \"certified-operators-jj2k9\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.632069 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.654450 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9xxr4"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.655395 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.657873 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.658143 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-utilities\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.658199 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-catalog-content\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.658263 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc48v\" (UniqueName: \"kubernetes.io/projected/3acd7391-9434-4efe-9526-ba40938af59e-kube-api-access-mc48v\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.658426 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.158403965 +0000 UTC m=+150.047994775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.678000 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v8gbl" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768375 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768415 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7vrj\" (UniqueName: \"kubernetes.io/projected/56994aba-f196-4e09-a51a-6053d114a640-kube-api-access-g7vrj\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768447 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9xxr4"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768482 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-utilities\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768522 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-catalog-content\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768551 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc48v\" (UniqueName: \"kubernetes.io/projected/3acd7391-9434-4efe-9526-ba40938af59e-kube-api-access-mc48v\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768586 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-catalog-content\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.768609 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-utilities\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.768746 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.268728797 +0000 UTC m=+150.158319677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.769000 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-utilities\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.769230 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-catalog-content\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.808282 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc48v\" (UniqueName: \"kubernetes.io/projected/3acd7391-9434-4efe-9526-ba40938af59e-kube-api-access-mc48v\") pod \"community-operators-w77l6\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.809926 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:05 crc kubenswrapper[4706]: W1127 07:11:05.869517 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-cffc09a83563a15a726ff22de2fcb5d7e559a04596b14c3ea83555932a1bb875 WatchSource:0}: Error finding container cffc09a83563a15a726ff22de2fcb5d7e559a04596b14c3ea83555932a1bb875: Status 404 returned error can't find the container with id cffc09a83563a15a726ff22de2fcb5d7e559a04596b14c3ea83555932a1bb875 Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.869869 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.369831785 +0000 UTC m=+150.259422595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.869788 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.870453 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-catalog-content\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.870498 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-utilities\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.870537 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.870561 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7vrj\" (UniqueName: \"kubernetes.io/projected/56994aba-f196-4e09-a51a-6053d114a640-kube-api-access-g7vrj\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.871052 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-catalog-content\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.871183 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-utilities\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.871379 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.37136313 +0000 UTC m=+150.260953940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.912557 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7vrj\" (UniqueName: \"kubernetes.io/projected/56994aba-f196-4e09-a51a-6053d114a640-kube-api-access-g7vrj\") pod \"certified-operators-9xxr4\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.917722 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.918510 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.926062 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.926406 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.937540 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 07:11:05 crc kubenswrapper[4706]: I1127 07:11:05.971397 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:05 crc kubenswrapper[4706]: E1127 07:11:05.971781 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.471749532 +0000 UTC m=+150.361340342 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.028420 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.064642 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:06 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:06 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:06 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.065051 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.073799 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.073849 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.073892 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.074276 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.574260621 +0000 UTC m=+150.463851431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.176167 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.176708 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.176824 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.176965 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.676923855 +0000 UTC m=+150.566514665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.177094 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.201916 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.208553 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tq6jq"] Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.257796 4706 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.276182 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.278188 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.278551 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.778537311 +0000 UTC m=+150.668128121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.288474 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.296343 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj2k9"] Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.377701 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w77l6"] Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.379703 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2bg9\" (UniqueName: \"kubernetes.io/projected/a709c580-4d9f-457d-a6cf-67a2831f4dcd-kube-api-access-h2bg9\") pod \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.379762 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a709c580-4d9f-457d-a6cf-67a2831f4dcd-secret-volume\") pod \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.379805 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume\") pod \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\" (UID: \"a709c580-4d9f-457d-a6cf-67a2831f4dcd\") " Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.379923 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.380339 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.880319903 +0000 UTC m=+150.769910703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.384372 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume" (OuterVolumeSpecName: "config-volume") pod "a709c580-4d9f-457d-a6cf-67a2831f4dcd" (UID: "a709c580-4d9f-457d-a6cf-67a2831f4dcd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.388029 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a709c580-4d9f-457d-a6cf-67a2831f4dcd-kube-api-access-h2bg9" (OuterVolumeSpecName: "kube-api-access-h2bg9") pod "a709c580-4d9f-457d-a6cf-67a2831f4dcd" (UID: "a709c580-4d9f-457d-a6cf-67a2831f4dcd"). InnerVolumeSpecName "kube-api-access-h2bg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.388939 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a709c580-4d9f-457d-a6cf-67a2831f4dcd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a709c580-4d9f-457d-a6cf-67a2831f4dcd" (UID: "a709c580-4d9f-457d-a6cf-67a2831f4dcd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.477626 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9xxr4"] Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.482422 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.482491 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a709c580-4d9f-457d-a6cf-67a2831f4dcd-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.482504 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2bg9\" (UniqueName: \"kubernetes.io/projected/a709c580-4d9f-457d-a6cf-67a2831f4dcd-kube-api-access-h2bg9\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.482514 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a709c580-4d9f-457d-a6cf-67a2831f4dcd-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.482753 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:06.982740179 +0000 UTC m=+150.872330989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.485873 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ca4ad779365e0a8b587e8ab35c35648582aa6d43485f8801a7ef708d929a8aec"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.485915 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7cf698039c748b41f252c5192ba00c8cf500b997ffebc9416d9ede673c57e2cf"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.519607 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" event={"ID":"a56eee0b-d1b8-47e0-887f-4fe2d2ed47df","Type":"ContainerStarted","Data":"f494b0a9796b8eaef79a568139ceb8b132af3f41ff7f754e1dab48f7c51e50c1"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.525714 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj2k9" event={"ID":"a8ac4a84-ddce-41e8-9da0-efd60117aa67","Type":"ContainerStarted","Data":"bfd342ab8d81c3d8368b0aecfcc292fb1e535870f2d51f160fa14a208a0488d4"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.557510 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-8jrds" podStartSLOduration=10.55749245 podStartE2EDuration="10.55749245s" podCreationTimestamp="2025-11-27 07:10:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:06.553728193 +0000 UTC m=+150.443318993" watchObservedRunningTime="2025-11-27 07:11:06.55749245 +0000 UTC m=+150.447083260" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.581441 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f89010e12ea1a7d2d829ea74699c60add30c61a442ae000171cf817e45813b80"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.581496 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cffc09a83563a15a726ff22de2fcb5d7e559a04596b14c3ea83555932a1bb875"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.582879 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.584401 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.584475 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.084458919 +0000 UTC m=+150.974049729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.594997 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" event={"ID":"a709c580-4d9f-457d-a6cf-67a2831f4dcd","Type":"ContainerDied","Data":"ef4fbf08c7d28422dd568b7a65b1586888423eb0271407db4a055074d3b1b97f"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.595041 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef4fbf08c7d28422dd568b7a65b1586888423eb0271407db4a055074d3b1b97f" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.595134 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403780-gsqzn" Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.596600 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w77l6" event={"ID":"3acd7391-9434-4efe-9526-ba40938af59e","Type":"ContainerStarted","Data":"cccf956716a7d176ab33b32ff7bbc33ddf5863f38c054dff54cb3f71016d5d94"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.597905 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"03ea28af4a56d1651e8f76b923c81a1015d14a26de73a07919ec1a5474ae8e6d"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.597935 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e8f869c224c484ab0d9f3fd1d595be6f6431c86c3c9fae6969dc00e1f83aa49e"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.599492 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6jq" event={"ID":"43c8fce5-14bd-4d14-b488-50600b596cb5","Type":"ContainerStarted","Data":"3b323d55f9be44626d90defc91a7c25f02545e59c9ab2f0c376d1828cd6851c8"} Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.686669 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.687770 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.187748925 +0000 UTC m=+151.077339735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.787321 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.787592 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.287558096 +0000 UTC m=+151.177148906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.787643 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.788094 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.288087945 +0000 UTC m=+151.177678755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.793865 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 07:11:06 crc kubenswrapper[4706]: W1127 07:11:06.839063 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3944bd9b_b6f3_4593_bc0c_c9b5410f8ac5.slice/crio-001caf319ac2f1f9a55a029f02275e407b184781a4ca7cc28d2cd0c556718846 WatchSource:0}: Error finding container 001caf319ac2f1f9a55a029f02275e407b184781a4ca7cc28d2cd0c556718846: Status 404 returned error can't find the container with id 001caf319ac2f1f9a55a029f02275e407b184781a4ca7cc28d2cd0c556718846 Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.888670 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.888921 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.388883501 +0000 UTC m=+151.278474311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.889466 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.890004 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.389992941 +0000 UTC m=+151.279583751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.990414 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.990622 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.49058838 +0000 UTC m=+151.380179200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:06 crc kubenswrapper[4706]: I1127 07:11:06.990822 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:06 crc kubenswrapper[4706]: E1127 07:11:06.991152 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.491142191 +0000 UTC m=+151.380733001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.029278 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:07 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:07 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:07 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.029354 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.091714 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:07 crc kubenswrapper[4706]: E1127 07:11:07.091938 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.591907736 +0000 UTC m=+151.481498546 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.092035 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:07 crc kubenswrapper[4706]: E1127 07:11:07.092406 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 07:11:07.592392424 +0000 UTC m=+151.481983234 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2gpf6" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.116681 4706 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-27T07:11:06.258206653Z","Handler":null,"Name":""} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.121057 4706 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.121102 4706 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.193060 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.196970 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.250116 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gxm5z"] Nov 27 07:11:07 crc kubenswrapper[4706]: E1127 07:11:07.250609 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a709c580-4d9f-457d-a6cf-67a2831f4dcd" containerName="collect-profiles" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.250627 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a709c580-4d9f-457d-a6cf-67a2831f4dcd" containerName="collect-profiles" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.250798 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a709c580-4d9f-457d-a6cf-67a2831f4dcd" containerName="collect-profiles" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.251918 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.253845 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.260740 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxm5z"] Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.294490 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.297652 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.297689 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.321267 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2gpf6\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.395983 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-catalog-content\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.396037 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-utilities\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.396067 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j74fc\" (UniqueName: \"kubernetes.io/projected/3b1176cb-10d1-4516-9ece-d51bb6ca515d-kube-api-access-j74fc\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.414933 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.497121 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-catalog-content\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.497538 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-utilities\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.497571 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j74fc\" (UniqueName: \"kubernetes.io/projected/3b1176cb-10d1-4516-9ece-d51bb6ca515d-kube-api-access-j74fc\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.497974 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-catalog-content\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.498027 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-utilities\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.520244 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j74fc\" (UniqueName: \"kubernetes.io/projected/3b1176cb-10d1-4516-9ece-d51bb6ca515d-kube-api-access-j74fc\") pod \"redhat-marketplace-gxm5z\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.568676 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.612237 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5","Type":"ContainerStarted","Data":"c7c6ac87e3a8157786a3e0237b00e4e320ef9895718be30112cbbe3dc674cc99"} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.612278 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5","Type":"ContainerStarted","Data":"001caf319ac2f1f9a55a029f02275e407b184781a4ca7cc28d2cd0c556718846"} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.614078 4706 generic.go:334] "Generic (PLEG): container finished" podID="3acd7391-9434-4efe-9526-ba40938af59e" containerID="378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb" exitCode=0 Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.614116 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w77l6" event={"ID":"3acd7391-9434-4efe-9526-ba40938af59e","Type":"ContainerDied","Data":"378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb"} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.615659 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.617066 4706 generic.go:334] "Generic (PLEG): container finished" podID="56994aba-f196-4e09-a51a-6053d114a640" containerID="b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39" exitCode=0 Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.617139 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xxr4" event={"ID":"56994aba-f196-4e09-a51a-6053d114a640","Type":"ContainerDied","Data":"b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39"} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.617165 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xxr4" event={"ID":"56994aba-f196-4e09-a51a-6053d114a640","Type":"ContainerStarted","Data":"a4ae9247c6ddad871d68505a7bd116c906a368f66c477d8782bb0e4da5b980e8"} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.620986 4706 generic.go:334] "Generic (PLEG): container finished" podID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerID="00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75" exitCode=0 Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.621040 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6jq" event={"ID":"43c8fce5-14bd-4d14-b488-50600b596cb5","Type":"ContainerDied","Data":"00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75"} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.625281 4706 generic.go:334] "Generic (PLEG): container finished" podID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerID="8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f" exitCode=0 Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.625373 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj2k9" event={"ID":"a8ac4a84-ddce-41e8-9da0-efd60117aa67","Type":"ContainerDied","Data":"8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f"} Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.644607 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2gpf6"] Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.648584 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.648564579 podStartE2EDuration="2.648564579s" podCreationTimestamp="2025-11-27 07:11:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:07.635838967 +0000 UTC m=+151.525429777" watchObservedRunningTime="2025-11-27 07:11:07.648564579 +0000 UTC m=+151.538155389" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.664394 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dmvlr"] Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.665468 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.685627 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmvlr"] Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.703809 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-catalog-content\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.704347 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-utilities\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.704406 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwhm\" (UniqueName: \"kubernetes.io/projected/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-kube-api-access-jpwhm\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.805383 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwhm\" (UniqueName: \"kubernetes.io/projected/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-kube-api-access-jpwhm\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.805647 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-catalog-content\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.805718 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-utilities\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.810930 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-catalog-content\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.812885 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-utilities\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: I1127 07:11:07.846032 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwhm\" (UniqueName: \"kubernetes.io/projected/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-kube-api-access-jpwhm\") pod \"redhat-marketplace-dmvlr\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:07 crc kubenswrapper[4706]: E1127 07:11:07.920499 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod3944bd9b_b6f3_4593_bc0c_c9b5410f8ac5.slice/crio-c7c6ac87e3a8157786a3e0237b00e4e320ef9895718be30112cbbe3dc674cc99.scope\": RecentStats: unable to find data in memory cache]" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.004032 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.028357 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:08 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:08 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:08 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.028431 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.153268 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxm5z"] Nov 27 07:11:08 crc kubenswrapper[4706]: W1127 07:11:08.169013 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b1176cb_10d1_4516_9ece_d51bb6ca515d.slice/crio-ce4ba0714eecf1b51e998693d3716fa7a26cf3aea07ff358c5828944aaf3f325 WatchSource:0}: Error finding container ce4ba0714eecf1b51e998693d3716fa7a26cf3aea07ff358c5828944aaf3f325: Status 404 returned error can't find the container with id ce4ba0714eecf1b51e998693d3716fa7a26cf3aea07ff358c5828944aaf3f325 Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.209021 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmvlr"] Nov 27 07:11:08 crc kubenswrapper[4706]: W1127 07:11:08.233827 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4e1e493_c2ef_42fd_b6ca_41c9bc0cff69.slice/crio-58b2f54fc06ef5dfb3d20f3ac8257acc752ec2d0c585fe5c5322033db099a0f6 WatchSource:0}: Error finding container 58b2f54fc06ef5dfb3d20f3ac8257acc752ec2d0c585fe5c5322033db099a0f6: Status 404 returned error can't find the container with id 58b2f54fc06ef5dfb3d20f3ac8257acc752ec2d0c585fe5c5322033db099a0f6 Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.255682 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8vv7v"] Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.256744 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.258846 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.268475 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8vv7v"] Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.298799 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-dh98x container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.299333 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dh98x" podUID="094c9ba3-fb86-48e5-93e0-f5f1f55cc12e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.298848 4706 patch_prober.go:28] interesting pod/downloads-7954f5f757-dh98x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.299447 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dh98x" podUID="094c9ba3-fb86-48e5-93e0-f5f1f55cc12e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.312357 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-utilities\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.312454 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2llrs\" (UniqueName: \"kubernetes.io/projected/adf72c72-f7cf-4ed2-9356-ef2c880d8164-kube-api-access-2llrs\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.312484 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-catalog-content\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.414143 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2llrs\" (UniqueName: \"kubernetes.io/projected/adf72c72-f7cf-4ed2-9356-ef2c880d8164-kube-api-access-2llrs\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.414255 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-catalog-content\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.414311 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-utilities\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.415180 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-utilities\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.415634 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-catalog-content\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.437175 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2llrs\" (UniqueName: \"kubernetes.io/projected/adf72c72-f7cf-4ed2-9356-ef2c880d8164-kube-api-access-2llrs\") pod \"redhat-operators-8vv7v\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.606579 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.610940 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.612030 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-nhj5d" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.623903 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.625134 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.651733 4706 patch_prober.go:28] interesting pod/console-f9d7485db-wt6br container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.651792 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wt6br" podUID="d8e36cad-dbdd-458c-9862-7131d620ba34" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.653174 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tnxnn"] Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.655453 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.680698 4706 generic.go:334] "Generic (PLEG): container finished" podID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerID="a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4" exitCode=0 Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.680798 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmvlr" event={"ID":"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69","Type":"ContainerDied","Data":"a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4"} Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.680832 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmvlr" event={"ID":"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69","Type":"ContainerStarted","Data":"58b2f54fc06ef5dfb3d20f3ac8257acc752ec2d0c585fe5c5322033db099a0f6"} Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.700521 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tnxnn"] Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.713979 4706 generic.go:334] "Generic (PLEG): container finished" podID="3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5" containerID="c7c6ac87e3a8157786a3e0237b00e4e320ef9895718be30112cbbe3dc674cc99" exitCode=0 Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.714159 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5","Type":"ContainerDied","Data":"c7c6ac87e3a8157786a3e0237b00e4e320ef9895718be30112cbbe3dc674cc99"} Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.719598 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-catalog-content\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.719703 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcpm4\" (UniqueName: \"kubernetes.io/projected/24597f11-1294-4631-82db-4a62bad47e8c-kube-api-access-lcpm4\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.720415 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-utilities\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.730023 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" event={"ID":"bdec2b7f-560f-44e9-b665-2ff892495d19","Type":"ContainerStarted","Data":"5918a2b15e0e21b34613f46b8f0cdabf49ad15328d82288b5c58349bbe7b2729"} Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.730105 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" event={"ID":"bdec2b7f-560f-44e9-b665-2ff892495d19","Type":"ContainerStarted","Data":"163909d20be1a0aa6de660b3c2484691dc632973bb1e9234234997dacb42a10e"} Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.731713 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.765340 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerID="2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d" exitCode=0 Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.766522 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxm5z" event={"ID":"3b1176cb-10d1-4516-9ece-d51bb6ca515d","Type":"ContainerDied","Data":"2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d"} Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.766551 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxm5z" event={"ID":"3b1176cb-10d1-4516-9ece-d51bb6ca515d","Type":"ContainerStarted","Data":"ce4ba0714eecf1b51e998693d3716fa7a26cf3aea07ff358c5828944aaf3f325"} Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.811988 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.813404 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.813563 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-974lt" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.821785 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nc99l" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.829710 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-catalog-content\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.829783 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcpm4\" (UniqueName: \"kubernetes.io/projected/24597f11-1294-4631-82db-4a62bad47e8c-kube-api-access-lcpm4\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.829967 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-utilities\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.831812 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-catalog-content\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.833520 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-utilities\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.853843 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.859718 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcpm4\" (UniqueName: \"kubernetes.io/projected/24597f11-1294-4631-82db-4a62bad47e8c-kube-api-access-lcpm4\") pod \"redhat-operators-tnxnn\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:08 crc kubenswrapper[4706]: I1127 07:11:08.870189 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" podStartSLOduration=131.870162773 podStartE2EDuration="2m11.870162773s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:08.868574396 +0000 UTC m=+152.758165206" watchObservedRunningTime="2025-11-27 07:11:08.870162773 +0000 UTC m=+152.759753583" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.025355 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.029489 4706 patch_prober.go:28] interesting pod/router-default-5444994796-8bwv8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 07:11:09 crc kubenswrapper[4706]: [-]has-synced failed: reason withheld Nov 27 07:11:09 crc kubenswrapper[4706]: [+]process-running ok Nov 27 07:11:09 crc kubenswrapper[4706]: healthz check failed Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.029535 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8bwv8" podUID="776f96c1-efaa-4cdd-9ef4-0533af4f2ca0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.030074 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.309630 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8vv7v"] Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.344863 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tnxnn"] Nov 27 07:11:09 crc kubenswrapper[4706]: W1127 07:11:09.370173 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24597f11_1294_4631_82db_4a62bad47e8c.slice/crio-7ce56a908c327b0dc78b94207562540b50d7d1db9757c94b4575a7545e4f8e44 WatchSource:0}: Error finding container 7ce56a908c327b0dc78b94207562540b50d7d1db9757c94b4575a7545e4f8e44: Status 404 returned error can't find the container with id 7ce56a908c327b0dc78b94207562540b50d7d1db9757c94b4575a7545e4f8e44 Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.449517 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.450296 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.453385 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.453568 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.456858 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.554286 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07979884-f807-4ca2-ae37-5c5239f77ea4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.555014 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07979884-f807-4ca2-ae37-5c5239f77ea4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.656718 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07979884-f807-4ca2-ae37-5c5239f77ea4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.656784 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07979884-f807-4ca2-ae37-5c5239f77ea4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.656903 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07979884-f807-4ca2-ae37-5c5239f77ea4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.680809 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07979884-f807-4ca2-ae37-5c5239f77ea4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.784166 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.789553 4706 generic.go:334] "Generic (PLEG): container finished" podID="24597f11-1294-4631-82db-4a62bad47e8c" containerID="ee9cddecdab2baa44a4990bfdcce59398f66cfc9f06825deb2fcd36ff95acc1f" exitCode=0 Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.789683 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnxnn" event={"ID":"24597f11-1294-4631-82db-4a62bad47e8c","Type":"ContainerDied","Data":"ee9cddecdab2baa44a4990bfdcce59398f66cfc9f06825deb2fcd36ff95acc1f"} Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.789764 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnxnn" event={"ID":"24597f11-1294-4631-82db-4a62bad47e8c","Type":"ContainerStarted","Data":"7ce56a908c327b0dc78b94207562540b50d7d1db9757c94b4575a7545e4f8e44"} Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.816080 4706 generic.go:334] "Generic (PLEG): container finished" podID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerID="28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e" exitCode=0 Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.816730 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8vv7v" event={"ID":"adf72c72-f7cf-4ed2-9356-ef2c880d8164","Type":"ContainerDied","Data":"28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e"} Nov 27 07:11:09 crc kubenswrapper[4706]: I1127 07:11:09.816805 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8vv7v" event={"ID":"adf72c72-f7cf-4ed2-9356-ef2c880d8164","Type":"ContainerStarted","Data":"42aac8c7c45033838425b5187b8e88e37a15dfee873c45ae2003256a6f2686c3"} Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.034805 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.038288 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-8bwv8" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.078738 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.163869 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kubelet-dir\") pod \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.163996 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kube-api-access\") pod \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\" (UID: \"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5\") " Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.164072 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5" (UID: "3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.164430 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.179445 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5" (UID: "3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.265802 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.276270 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 07:11:10 crc kubenswrapper[4706]: W1127 07:11:10.326460 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod07979884_f807_4ca2_ae37_5c5239f77ea4.slice/crio-5980eb8d23690ceea5e32d212479ff667d1385f898725832b0560491e0424b5a WatchSource:0}: Error finding container 5980eb8d23690ceea5e32d212479ff667d1385f898725832b0560491e0424b5a: Status 404 returned error can't find the container with id 5980eb8d23690ceea5e32d212479ff667d1385f898725832b0560491e0424b5a Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.825062 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"07979884-f807-4ca2-ae37-5c5239f77ea4","Type":"ContainerStarted","Data":"5980eb8d23690ceea5e32d212479ff667d1385f898725832b0560491e0424b5a"} Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.834187 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.834534 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5","Type":"ContainerDied","Data":"001caf319ac2f1f9a55a029f02275e407b184781a4ca7cc28d2cd0c556718846"} Nov 27 07:11:10 crc kubenswrapper[4706]: I1127 07:11:10.834559 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="001caf319ac2f1f9a55a029f02275e407b184781a4ca7cc28d2cd0c556718846" Nov 27 07:11:11 crc kubenswrapper[4706]: I1127 07:11:11.172185 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-v9mq5" Nov 27 07:11:11 crc kubenswrapper[4706]: I1127 07:11:11.844136 4706 generic.go:334] "Generic (PLEG): container finished" podID="07979884-f807-4ca2-ae37-5c5239f77ea4" containerID="ffa9cfd409eb41c3f318d118581b47fe0e4f7caa45fd744de5cb2a033e7da56b" exitCode=0 Nov 27 07:11:11 crc kubenswrapper[4706]: I1127 07:11:11.844195 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"07979884-f807-4ca2-ae37-5c5239f77ea4","Type":"ContainerDied","Data":"ffa9cfd409eb41c3f318d118581b47fe0e4f7caa45fd744de5cb2a033e7da56b"} Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.147404 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.217281 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07979884-f807-4ca2-ae37-5c5239f77ea4-kube-api-access\") pod \"07979884-f807-4ca2-ae37-5c5239f77ea4\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.217367 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07979884-f807-4ca2-ae37-5c5239f77ea4-kubelet-dir\") pod \"07979884-f807-4ca2-ae37-5c5239f77ea4\" (UID: \"07979884-f807-4ca2-ae37-5c5239f77ea4\") " Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.217486 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07979884-f807-4ca2-ae37-5c5239f77ea4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "07979884-f807-4ca2-ae37-5c5239f77ea4" (UID: "07979884-f807-4ca2-ae37-5c5239f77ea4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.217824 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07979884-f807-4ca2-ae37-5c5239f77ea4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.233458 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07979884-f807-4ca2-ae37-5c5239f77ea4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "07979884-f807-4ca2-ae37-5c5239f77ea4" (UID: "07979884-f807-4ca2-ae37-5c5239f77ea4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.318982 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07979884-f807-4ca2-ae37-5c5239f77ea4-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.875880 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"07979884-f807-4ca2-ae37-5c5239f77ea4","Type":"ContainerDied","Data":"5980eb8d23690ceea5e32d212479ff667d1385f898725832b0560491e0424b5a"} Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.876207 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5980eb8d23690ceea5e32d212479ff667d1385f898725832b0560491e0424b5a" Nov 27 07:11:13 crc kubenswrapper[4706]: I1127 07:11:13.875997 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 07:11:15 crc kubenswrapper[4706]: I1127 07:11:15.178495 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:11:15 crc kubenswrapper[4706]: I1127 07:11:15.178576 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:11:18 crc kubenswrapper[4706]: I1127 07:11:18.304344 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-dh98x" Nov 27 07:11:18 crc kubenswrapper[4706]: I1127 07:11:18.628074 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:11:18 crc kubenswrapper[4706]: I1127 07:11:18.633716 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wt6br" Nov 27 07:11:19 crc kubenswrapper[4706]: I1127 07:11:19.922445 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:11:19 crc kubenswrapper[4706]: I1127 07:11:19.929125 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab5bf95b-5c0d-4804-bc97-43b79f6f2fea-metrics-certs\") pod \"network-metrics-daemon-zxp2h\" (UID: \"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea\") " pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:11:20 crc kubenswrapper[4706]: I1127 07:11:20.199298 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zxp2h" Nov 27 07:11:27 crc kubenswrapper[4706]: I1127 07:11:27.420346 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:11:32 crc kubenswrapper[4706]: I1127 07:11:32.082288 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zxp2h"] Nov 27 07:11:36 crc kubenswrapper[4706]: E1127 07:11:36.041329 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 27 07:11:36 crc kubenswrapper[4706]: E1127 07:11:36.042821 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mc48v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-w77l6_openshift-marketplace(3acd7391-9434-4efe-9526-ba40938af59e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 07:11:36 crc kubenswrapper[4706]: E1127 07:11:36.044926 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-w77l6" podUID="3acd7391-9434-4efe-9526-ba40938af59e" Nov 27 07:11:36 crc kubenswrapper[4706]: E1127 07:11:36.047571 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 27 07:11:36 crc kubenswrapper[4706]: E1127 07:11:36.047721 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l5d64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tq6jq_openshift-marketplace(43c8fce5-14bd-4d14-b488-50600b596cb5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 07:11:36 crc kubenswrapper[4706]: E1127 07:11:36.049581 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tq6jq" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" Nov 27 07:11:36 crc kubenswrapper[4706]: W1127 07:11:36.973818 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab5bf95b_5c0d_4804_bc97_43b79f6f2fea.slice/crio-50cd588f0dc7d61cf3b5e5edbf2f78bd710d265ac5c6135d720ab717c0617b58 WatchSource:0}: Error finding container 50cd588f0dc7d61cf3b5e5edbf2f78bd710d265ac5c6135d720ab717c0617b58: Status 404 returned error can't find the container with id 50cd588f0dc7d61cf3b5e5edbf2f78bd710d265ac5c6135d720ab717c0617b58 Nov 27 07:11:37 crc kubenswrapper[4706]: I1127 07:11:37.000859 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" event={"ID":"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea","Type":"ContainerStarted","Data":"50cd588f0dc7d61cf3b5e5edbf2f78bd710d265ac5c6135d720ab717c0617b58"} Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.009365 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w77l6" podUID="3acd7391-9434-4efe-9526-ba40938af59e" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.009392 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tq6jq" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.025581 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.025804 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jpwhm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dmvlr_openshift-marketplace(b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.027837 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dmvlr" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.048602 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.048733 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j74fc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-gxm5z_openshift-marketplace(3b1176cb-10d1-4516-9ece-d51bb6ca515d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.049940 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-gxm5z" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.086107 4706 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.086259 4706 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lcpm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tnxnn_openshift-marketplace(24597f11-1294-4631-82db-4a62bad47e8c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 07:11:37 crc kubenswrapper[4706]: E1127 07:11:37.087594 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-tnxnn" podUID="24597f11-1294-4631-82db-4a62bad47e8c" Nov 27 07:11:37 crc kubenswrapper[4706]: I1127 07:11:37.335212 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fqglc"] Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.006416 4706 generic.go:334] "Generic (PLEG): container finished" podID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerID="9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c" exitCode=0 Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.006503 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8vv7v" event={"ID":"adf72c72-f7cf-4ed2-9356-ef2c880d8164","Type":"ContainerDied","Data":"9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c"} Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.007885 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" event={"ID":"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea","Type":"ContainerStarted","Data":"708ea50446e5231ca4ce04eaa07a29fba862aaf54f90f132d517de43355edd7b"} Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.007921 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zxp2h" event={"ID":"ab5bf95b-5c0d-4804-bc97-43b79f6f2fea","Type":"ContainerStarted","Data":"64be5bd89f1a6158393c3df31a4b1b69eede9a036fabef3ca08722e04951ca25"} Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.009567 4706 generic.go:334] "Generic (PLEG): container finished" podID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerID="89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785" exitCode=0 Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.009623 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj2k9" event={"ID":"a8ac4a84-ddce-41e8-9da0-efd60117aa67","Type":"ContainerDied","Data":"89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785"} Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.013324 4706 generic.go:334] "Generic (PLEG): container finished" podID="56994aba-f196-4e09-a51a-6053d114a640" containerID="9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69" exitCode=0 Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.013417 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xxr4" event={"ID":"56994aba-f196-4e09-a51a-6053d114a640","Type":"ContainerDied","Data":"9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69"} Nov 27 07:11:38 crc kubenswrapper[4706]: E1127 07:11:38.015324 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-gxm5z" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" Nov 27 07:11:38 crc kubenswrapper[4706]: E1127 07:11:38.015420 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dmvlr" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" Nov 27 07:11:38 crc kubenswrapper[4706]: E1127 07:11:38.019879 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tnxnn" podUID="24597f11-1294-4631-82db-4a62bad47e8c" Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.042280 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zxp2h" podStartSLOduration=161.042264958 podStartE2EDuration="2m41.042264958s" podCreationTimestamp="2025-11-27 07:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:38.03857528 +0000 UTC m=+181.928166100" watchObservedRunningTime="2025-11-27 07:11:38.042264958 +0000 UTC m=+181.931855758" Nov 27 07:11:38 crc kubenswrapper[4706]: I1127 07:11:38.783765 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gdp66" Nov 27 07:11:40 crc kubenswrapper[4706]: I1127 07:11:40.024842 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj2k9" event={"ID":"a8ac4a84-ddce-41e8-9da0-efd60117aa67","Type":"ContainerStarted","Data":"0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65"} Nov 27 07:11:40 crc kubenswrapper[4706]: I1127 07:11:40.027250 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xxr4" event={"ID":"56994aba-f196-4e09-a51a-6053d114a640","Type":"ContainerStarted","Data":"b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c"} Nov 27 07:11:40 crc kubenswrapper[4706]: I1127 07:11:40.029345 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8vv7v" event={"ID":"adf72c72-f7cf-4ed2-9356-ef2c880d8164","Type":"ContainerStarted","Data":"a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f"} Nov 27 07:11:40 crc kubenswrapper[4706]: I1127 07:11:40.054406 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jj2k9" podStartSLOduration=3.632781908 podStartE2EDuration="35.054388702s" podCreationTimestamp="2025-11-27 07:11:05 +0000 UTC" firstStartedPulling="2025-11-27 07:11:07.634720646 +0000 UTC m=+151.524311456" lastFinishedPulling="2025-11-27 07:11:39.05632744 +0000 UTC m=+182.945918250" observedRunningTime="2025-11-27 07:11:40.051838601 +0000 UTC m=+183.941429411" watchObservedRunningTime="2025-11-27 07:11:40.054388702 +0000 UTC m=+183.943979512" Nov 27 07:11:40 crc kubenswrapper[4706]: I1127 07:11:40.077160 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9xxr4" podStartSLOduration=3.718356143 podStartE2EDuration="35.077143956s" podCreationTimestamp="2025-11-27 07:11:05 +0000 UTC" firstStartedPulling="2025-11-27 07:11:07.619327848 +0000 UTC m=+151.508918658" lastFinishedPulling="2025-11-27 07:11:38.978115661 +0000 UTC m=+182.867706471" observedRunningTime="2025-11-27 07:11:40.075926337 +0000 UTC m=+183.965517147" watchObservedRunningTime="2025-11-27 07:11:40.077143956 +0000 UTC m=+183.966734756" Nov 27 07:11:40 crc kubenswrapper[4706]: I1127 07:11:40.094127 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8vv7v" podStartSLOduration=2.7995526120000003 podStartE2EDuration="32.094108426s" podCreationTimestamp="2025-11-27 07:11:08 +0000 UTC" firstStartedPulling="2025-11-27 07:11:09.818068328 +0000 UTC m=+153.707659138" lastFinishedPulling="2025-11-27 07:11:39.112624142 +0000 UTC m=+183.002214952" observedRunningTime="2025-11-27 07:11:40.092576298 +0000 UTC m=+183.982167108" watchObservedRunningTime="2025-11-27 07:11:40.094108426 +0000 UTC m=+183.983699236" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.247734 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 07:11:44 crc kubenswrapper[4706]: E1127 07:11:44.248322 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5" containerName="pruner" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.248339 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5" containerName="pruner" Nov 27 07:11:44 crc kubenswrapper[4706]: E1127 07:11:44.248362 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07979884-f807-4ca2-ae37-5c5239f77ea4" containerName="pruner" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.248370 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="07979884-f807-4ca2-ae37-5c5239f77ea4" containerName="pruner" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.248495 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="07979884-f807-4ca2-ae37-5c5239f77ea4" containerName="pruner" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.248509 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3944bd9b-b6f3-4593-bc0c-c9b5410f8ac5" containerName="pruner" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.248941 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.253743 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.253965 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.261464 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.276551 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.276632 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.378290 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.378372 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.378399 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.395149 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.571013 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.773042 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 07:11:44 crc kubenswrapper[4706]: I1127 07:11:44.817313 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 07:11:45 crc kubenswrapper[4706]: I1127 07:11:45.053971 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"357a79d7-fc2e-4c6e-bfb3-835da7a3deec","Type":"ContainerStarted","Data":"b578928f530e46005c8211771318efbf405ac8ef2508d22164bc002e0c811726"} Nov 27 07:11:45 crc kubenswrapper[4706]: I1127 07:11:45.177884 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:11:45 crc kubenswrapper[4706]: I1127 07:11:45.177943 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:11:45 crc kubenswrapper[4706]: I1127 07:11:45.633058 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:45 crc kubenswrapper[4706]: I1127 07:11:45.634308 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.029817 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.029872 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.060846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"357a79d7-fc2e-4c6e-bfb3-835da7a3deec","Type":"ContainerStarted","Data":"3f752ea3e3b524219796c33ee1d4e27b7c2e3cf459cec6a2b996567194c15929"} Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.085282 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.08526343 podStartE2EDuration="2.08526343s" podCreationTimestamp="2025-11-27 07:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:46.083032849 +0000 UTC m=+189.972623659" watchObservedRunningTime="2025-11-27 07:11:46.08526343 +0000 UTC m=+189.974854240" Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.102513 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.102742 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.142574 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:11:46 crc kubenswrapper[4706]: I1127 07:11:46.147367 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:47 crc kubenswrapper[4706]: I1127 07:11:47.065808 4706 generic.go:334] "Generic (PLEG): container finished" podID="357a79d7-fc2e-4c6e-bfb3-835da7a3deec" containerID="3f752ea3e3b524219796c33ee1d4e27b7c2e3cf459cec6a2b996567194c15929" exitCode=0 Nov 27 07:11:47 crc kubenswrapper[4706]: I1127 07:11:47.065909 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"357a79d7-fc2e-4c6e-bfb3-835da7a3deec","Type":"ContainerDied","Data":"3f752ea3e3b524219796c33ee1d4e27b7c2e3cf459cec6a2b996567194c15929"} Nov 27 07:11:47 crc kubenswrapper[4706]: I1127 07:11:47.347385 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9xxr4"] Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.073303 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9xxr4" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="registry-server" containerID="cri-o://b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c" gracePeriod=2 Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.299284 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.393993 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.425597 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kubelet-dir\") pod \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.425660 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kube-api-access\") pod \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\" (UID: \"357a79d7-fc2e-4c6e-bfb3-835da7a3deec\") " Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.425877 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "357a79d7-fc2e-4c6e-bfb3-835da7a3deec" (UID: "357a79d7-fc2e-4c6e-bfb3-835da7a3deec"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.431141 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "357a79d7-fc2e-4c6e-bfb3-835da7a3deec" (UID: "357a79d7-fc2e-4c6e-bfb3-835da7a3deec"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.526873 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-catalog-content\") pod \"56994aba-f196-4e09-a51a-6053d114a640\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.526950 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7vrj\" (UniqueName: \"kubernetes.io/projected/56994aba-f196-4e09-a51a-6053d114a640-kube-api-access-g7vrj\") pod \"56994aba-f196-4e09-a51a-6053d114a640\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.526974 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-utilities\") pod \"56994aba-f196-4e09-a51a-6053d114a640\" (UID: \"56994aba-f196-4e09-a51a-6053d114a640\") " Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.527261 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.527278 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/357a79d7-fc2e-4c6e-bfb3-835da7a3deec-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.527738 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-utilities" (OuterVolumeSpecName: "utilities") pod "56994aba-f196-4e09-a51a-6053d114a640" (UID: "56994aba-f196-4e09-a51a-6053d114a640"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.531429 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56994aba-f196-4e09-a51a-6053d114a640-kube-api-access-g7vrj" (OuterVolumeSpecName: "kube-api-access-g7vrj") pod "56994aba-f196-4e09-a51a-6053d114a640" (UID: "56994aba-f196-4e09-a51a-6053d114a640"). InnerVolumeSpecName "kube-api-access-g7vrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.572329 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56994aba-f196-4e09-a51a-6053d114a640" (UID: "56994aba-f196-4e09-a51a-6053d114a640"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.611934 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.612291 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.629143 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.629198 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7vrj\" (UniqueName: \"kubernetes.io/projected/56994aba-f196-4e09-a51a-6053d114a640-kube-api-access-g7vrj\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.629252 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56994aba-f196-4e09-a51a-6053d114a640-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:11:48 crc kubenswrapper[4706]: I1127 07:11:48.663828 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.079254 4706 generic.go:334] "Generic (PLEG): container finished" podID="56994aba-f196-4e09-a51a-6053d114a640" containerID="b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c" exitCode=0 Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.079345 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xxr4" event={"ID":"56994aba-f196-4e09-a51a-6053d114a640","Type":"ContainerDied","Data":"b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c"} Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.079369 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xxr4" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.079414 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xxr4" event={"ID":"56994aba-f196-4e09-a51a-6053d114a640","Type":"ContainerDied","Data":"a4ae9247c6ddad871d68505a7bd116c906a368f66c477d8782bb0e4da5b980e8"} Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.079434 4706 scope.go:117] "RemoveContainer" containerID="b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.081680 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.082018 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"357a79d7-fc2e-4c6e-bfb3-835da7a3deec","Type":"ContainerDied","Data":"b578928f530e46005c8211771318efbf405ac8ef2508d22164bc002e0c811726"} Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.082051 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b578928f530e46005c8211771318efbf405ac8ef2508d22164bc002e0c811726" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.100230 4706 scope.go:117] "RemoveContainer" containerID="9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.100893 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9xxr4"] Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.105421 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9xxr4"] Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.118513 4706 scope.go:117] "RemoveContainer" containerID="b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.121013 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.135490 4706 scope.go:117] "RemoveContainer" containerID="b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c" Nov 27 07:11:49 crc kubenswrapper[4706]: E1127 07:11:49.136273 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c\": container with ID starting with b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c not found: ID does not exist" containerID="b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.136317 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c"} err="failed to get container status \"b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c\": rpc error: code = NotFound desc = could not find container \"b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c\": container with ID starting with b2455cb1b2252182602265d08c83d5bbcccc4c245366f3585456e326c29e518c not found: ID does not exist" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.136367 4706 scope.go:117] "RemoveContainer" containerID="9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69" Nov 27 07:11:49 crc kubenswrapper[4706]: E1127 07:11:49.136979 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69\": container with ID starting with 9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69 not found: ID does not exist" containerID="9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.137033 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69"} err="failed to get container status \"9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69\": rpc error: code = NotFound desc = could not find container \"9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69\": container with ID starting with 9c1bedb80e08c772e0145787446e4df3b5fab0ffe95e77a36358d1b52ed50a69 not found: ID does not exist" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.137053 4706 scope.go:117] "RemoveContainer" containerID="b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39" Nov 27 07:11:49 crc kubenswrapper[4706]: E1127 07:11:49.137891 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39\": container with ID starting with b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39 not found: ID does not exist" containerID="b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39" Nov 27 07:11:49 crc kubenswrapper[4706]: I1127 07:11:49.137931 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39"} err="failed to get container status \"b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39\": rpc error: code = NotFound desc = could not find container \"b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39\": container with ID starting with b37e197dff89bb5dc46f986e0d51599ddcd21382f18735acc95b89866eb14b39 not found: ID does not exist" Nov 27 07:11:50 crc kubenswrapper[4706]: I1127 07:11:50.799547 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56994aba-f196-4e09-a51a-6053d114a640" path="/var/lib/kubelet/pods/56994aba-f196-4e09-a51a-6053d114a640/volumes" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.646955 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 07:11:51 crc kubenswrapper[4706]: E1127 07:11:51.647538 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="registry-server" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.647562 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="registry-server" Nov 27 07:11:51 crc kubenswrapper[4706]: E1127 07:11:51.647584 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="extract-utilities" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.647593 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="extract-utilities" Nov 27 07:11:51 crc kubenswrapper[4706]: E1127 07:11:51.647604 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="extract-content" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.647613 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="extract-content" Nov 27 07:11:51 crc kubenswrapper[4706]: E1127 07:11:51.647627 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357a79d7-fc2e-4c6e-bfb3-835da7a3deec" containerName="pruner" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.647634 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="357a79d7-fc2e-4c6e-bfb3-835da7a3deec" containerName="pruner" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.647749 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="357a79d7-fc2e-4c6e-bfb3-835da7a3deec" containerName="pruner" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.647767 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="56994aba-f196-4e09-a51a-6053d114a640" containerName="registry-server" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.648199 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.650616 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.651206 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.661119 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.767058 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.767139 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-var-lock\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.767176 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e14e4ff4-0455-402f-8299-044856ac016d-kube-api-access\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.868728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.868803 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-var-lock\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.868840 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e14e4ff4-0455-402f-8299-044856ac016d-kube-api-access\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.869146 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.869179 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-var-lock\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.885126 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e14e4ff4-0455-402f-8299-044856ac016d-kube-api-access\") pod \"installer-9-crc\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:51 crc kubenswrapper[4706]: I1127 07:11:51.964526 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:11:52 crc kubenswrapper[4706]: I1127 07:11:52.106705 4706 generic.go:334] "Generic (PLEG): container finished" podID="3acd7391-9434-4efe-9526-ba40938af59e" containerID="a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76" exitCode=0 Nov 27 07:11:52 crc kubenswrapper[4706]: I1127 07:11:52.106970 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w77l6" event={"ID":"3acd7391-9434-4efe-9526-ba40938af59e","Type":"ContainerDied","Data":"a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76"} Nov 27 07:11:52 crc kubenswrapper[4706]: I1127 07:11:52.202430 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.119373 4706 generic.go:334] "Generic (PLEG): container finished" podID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerID="1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3" exitCode=0 Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.119430 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmvlr" event={"ID":"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69","Type":"ContainerDied","Data":"1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3"} Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.123677 4706 generic.go:334] "Generic (PLEG): container finished" podID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerID="0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98" exitCode=0 Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.123781 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6jq" event={"ID":"43c8fce5-14bd-4d14-b488-50600b596cb5","Type":"ContainerDied","Data":"0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98"} Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.129149 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w77l6" event={"ID":"3acd7391-9434-4efe-9526-ba40938af59e","Type":"ContainerStarted","Data":"b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02"} Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.130617 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e14e4ff4-0455-402f-8299-044856ac016d","Type":"ContainerStarted","Data":"9d76ea3d0e4555adfa880c6e49bb9a6506b75258ca9541ed2a1fc7a1b1d6065b"} Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.130652 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e14e4ff4-0455-402f-8299-044856ac016d","Type":"ContainerStarted","Data":"dfc6f07343f40b0868a81db13132310d3348e87f1223c03d829060370a677c5f"} Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.132475 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnxnn" event={"ID":"24597f11-1294-4631-82db-4a62bad47e8c","Type":"ContainerStarted","Data":"e990c1c678846daca31db721831bf84b68c2e8161115647e9b651e1485484cbb"} Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.151048 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.151033518 podStartE2EDuration="2.151033518s" podCreationTimestamp="2025-11-27 07:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:11:53.149002839 +0000 UTC m=+197.038593649" watchObservedRunningTime="2025-11-27 07:11:53.151033518 +0000 UTC m=+197.040624348" Nov 27 07:11:53 crc kubenswrapper[4706]: I1127 07:11:53.166603 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w77l6" podStartSLOduration=3.060168032 podStartE2EDuration="48.166587887s" podCreationTimestamp="2025-11-27 07:11:05 +0000 UTC" firstStartedPulling="2025-11-27 07:11:07.615371255 +0000 UTC m=+151.504962065" lastFinishedPulling="2025-11-27 07:11:52.72179111 +0000 UTC m=+196.611381920" observedRunningTime="2025-11-27 07:11:53.165714387 +0000 UTC m=+197.055305187" watchObservedRunningTime="2025-11-27 07:11:53.166587887 +0000 UTC m=+197.056178697" Nov 27 07:11:54 crc kubenswrapper[4706]: I1127 07:11:54.139554 4706 generic.go:334] "Generic (PLEG): container finished" podID="24597f11-1294-4631-82db-4a62bad47e8c" containerID="e990c1c678846daca31db721831bf84b68c2e8161115647e9b651e1485484cbb" exitCode=0 Nov 27 07:11:54 crc kubenswrapper[4706]: I1127 07:11:54.139629 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnxnn" event={"ID":"24597f11-1294-4631-82db-4a62bad47e8c","Type":"ContainerDied","Data":"e990c1c678846daca31db721831bf84b68c2e8161115647e9b651e1485484cbb"} Nov 27 07:11:54 crc kubenswrapper[4706]: I1127 07:11:54.142634 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmvlr" event={"ID":"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69","Type":"ContainerStarted","Data":"22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee"} Nov 27 07:11:54 crc kubenswrapper[4706]: I1127 07:11:54.144823 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6jq" event={"ID":"43c8fce5-14bd-4d14-b488-50600b596cb5","Type":"ContainerStarted","Data":"7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099"} Nov 27 07:11:54 crc kubenswrapper[4706]: I1127 07:11:54.176352 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tq6jq" podStartSLOduration=3.163586645 podStartE2EDuration="49.176335483s" podCreationTimestamp="2025-11-27 07:11:05 +0000 UTC" firstStartedPulling="2025-11-27 07:11:07.623990047 +0000 UTC m=+151.513580857" lastFinishedPulling="2025-11-27 07:11:53.636738885 +0000 UTC m=+197.526329695" observedRunningTime="2025-11-27 07:11:54.174197511 +0000 UTC m=+198.063788321" watchObservedRunningTime="2025-11-27 07:11:54.176335483 +0000 UTC m=+198.065926293" Nov 27 07:11:54 crc kubenswrapper[4706]: I1127 07:11:54.198136 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dmvlr" podStartSLOduration=2.25854819 podStartE2EDuration="47.198120983s" podCreationTimestamp="2025-11-27 07:11:07 +0000 UTC" firstStartedPulling="2025-11-27 07:11:08.685896079 +0000 UTC m=+152.575486889" lastFinishedPulling="2025-11-27 07:11:53.625468872 +0000 UTC m=+197.515059682" observedRunningTime="2025-11-27 07:11:54.194437218 +0000 UTC m=+198.084028028" watchObservedRunningTime="2025-11-27 07:11:54.198120983 +0000 UTC m=+198.087711793" Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.150272 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerID="53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab" exitCode=0 Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.150336 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxm5z" event={"ID":"3b1176cb-10d1-4516-9ece-d51bb6ca515d","Type":"ContainerDied","Data":"53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab"} Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.153350 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnxnn" event={"ID":"24597f11-1294-4631-82db-4a62bad47e8c","Type":"ContainerStarted","Data":"297311e6901b449baf19c5846251eade8677cb9b2ea832939bbbf599b3ded2ec"} Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.188773 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tnxnn" podStartSLOduration=2.403237504 podStartE2EDuration="47.18875683s" podCreationTimestamp="2025-11-27 07:11:08 +0000 UTC" firstStartedPulling="2025-11-27 07:11:09.813017655 +0000 UTC m=+153.702608465" lastFinishedPulling="2025-11-27 07:11:54.598536981 +0000 UTC m=+198.488127791" observedRunningTime="2025-11-27 07:11:55.186095299 +0000 UTC m=+199.075686119" watchObservedRunningTime="2025-11-27 07:11:55.18875683 +0000 UTC m=+199.078347640" Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.445820 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.445879 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.811020 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.811071 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:55 crc kubenswrapper[4706]: I1127 07:11:55.845934 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:11:56 crc kubenswrapper[4706]: I1127 07:11:56.162186 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxm5z" event={"ID":"3b1176cb-10d1-4516-9ece-d51bb6ca515d","Type":"ContainerStarted","Data":"35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52"} Nov 27 07:11:56 crc kubenswrapper[4706]: I1127 07:11:56.181716 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gxm5z" podStartSLOduration=2.425915686 podStartE2EDuration="49.181697136s" podCreationTimestamp="2025-11-27 07:11:07 +0000 UTC" firstStartedPulling="2025-11-27 07:11:08.7678146 +0000 UTC m=+152.657405410" lastFinishedPulling="2025-11-27 07:11:55.52359605 +0000 UTC m=+199.413186860" observedRunningTime="2025-11-27 07:11:56.177886806 +0000 UTC m=+200.067477616" watchObservedRunningTime="2025-11-27 07:11:56.181697136 +0000 UTC m=+200.071287946" Nov 27 07:11:56 crc kubenswrapper[4706]: I1127 07:11:56.484645 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tq6jq" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="registry-server" probeResult="failure" output=< Nov 27 07:11:56 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Nov 27 07:11:56 crc kubenswrapper[4706]: > Nov 27 07:11:57 crc kubenswrapper[4706]: I1127 07:11:57.570327 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:57 crc kubenswrapper[4706]: I1127 07:11:57.570384 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:57 crc kubenswrapper[4706]: I1127 07:11:57.613725 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:11:58 crc kubenswrapper[4706]: I1127 07:11:58.004662 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:58 crc kubenswrapper[4706]: I1127 07:11:58.004715 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:58 crc kubenswrapper[4706]: I1127 07:11:58.040925 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:58 crc kubenswrapper[4706]: I1127 07:11:58.206050 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:11:58 crc kubenswrapper[4706]: I1127 07:11:58.945500 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmvlr"] Nov 27 07:11:59 crc kubenswrapper[4706]: I1127 07:11:59.030880 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:11:59 crc kubenswrapper[4706]: I1127 07:11:59.030928 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.099473 4706 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tnxnn" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="registry-server" probeResult="failure" output=< Nov 27 07:12:00 crc kubenswrapper[4706]: timeout: failed to connect service ":50051" within 1s Nov 27 07:12:00 crc kubenswrapper[4706]: > Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.187104 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dmvlr" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="registry-server" containerID="cri-o://22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee" gracePeriod=2 Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.651788 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.680818 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-utilities\") pod \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.680914 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpwhm\" (UniqueName: \"kubernetes.io/projected/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-kube-api-access-jpwhm\") pod \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.680956 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-catalog-content\") pod \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\" (UID: \"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69\") " Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.681844 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-utilities" (OuterVolumeSpecName: "utilities") pod "b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" (UID: "b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.686072 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-kube-api-access-jpwhm" (OuterVolumeSpecName: "kube-api-access-jpwhm") pod "b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" (UID: "b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69"). InnerVolumeSpecName "kube-api-access-jpwhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.697415 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" (UID: "b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.782178 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.782433 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpwhm\" (UniqueName: \"kubernetes.io/projected/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-kube-api-access-jpwhm\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:00 crc kubenswrapper[4706]: I1127 07:12:00.782497 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.198004 4706 generic.go:334] "Generic (PLEG): container finished" podID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerID="22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee" exitCode=0 Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.198085 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmvlr" event={"ID":"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69","Type":"ContainerDied","Data":"22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee"} Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.198140 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmvlr" event={"ID":"b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69","Type":"ContainerDied","Data":"58b2f54fc06ef5dfb3d20f3ac8257acc752ec2d0c585fe5c5322033db099a0f6"} Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.198162 4706 scope.go:117] "RemoveContainer" containerID="22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.198403 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmvlr" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.221464 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmvlr"] Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.227623 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmvlr"] Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.228615 4706 scope.go:117] "RemoveContainer" containerID="1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.243438 4706 scope.go:117] "RemoveContainer" containerID="a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.268672 4706 scope.go:117] "RemoveContainer" containerID="22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee" Nov 27 07:12:01 crc kubenswrapper[4706]: E1127 07:12:01.269172 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee\": container with ID starting with 22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee not found: ID does not exist" containerID="22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.269346 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee"} err="failed to get container status \"22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee\": rpc error: code = NotFound desc = could not find container \"22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee\": container with ID starting with 22bb9177f90b1bd840d136ffd60791cf142da867f2a9ef731c5608e793c7a1ee not found: ID does not exist" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.269507 4706 scope.go:117] "RemoveContainer" containerID="1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3" Nov 27 07:12:01 crc kubenswrapper[4706]: E1127 07:12:01.270212 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3\": container with ID starting with 1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3 not found: ID does not exist" containerID="1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.270287 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3"} err="failed to get container status \"1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3\": rpc error: code = NotFound desc = could not find container \"1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3\": container with ID starting with 1b194c8c19a789478c323ebb91912a2fd400906984b333ad56d557e9a9dc87b3 not found: ID does not exist" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.270325 4706 scope.go:117] "RemoveContainer" containerID="a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4" Nov 27 07:12:01 crc kubenswrapper[4706]: E1127 07:12:01.270651 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4\": container with ID starting with a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4 not found: ID does not exist" containerID="a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4" Nov 27 07:12:01 crc kubenswrapper[4706]: I1127 07:12:01.270791 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4"} err="failed to get container status \"a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4\": rpc error: code = NotFound desc = could not find container \"a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4\": container with ID starting with a3b678b4d1f71a3015cb8883dcbae2ee7582be378e4870b867ebef4dc515e1f4 not found: ID does not exist" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.378263 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" podUID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" containerName="oauth-openshift" containerID="cri-o://8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e" gracePeriod=15 Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.727720 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.788700 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" path="/var/lib/kubelet/pods/b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69/volumes" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908213 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-idp-0-file-data\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908291 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-dir\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908317 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-policies\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908335 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908374 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-login\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908404 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-session\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908438 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-provider-selection\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908459 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-service-ca\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908478 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-router-certs\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908503 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-error\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908525 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-serving-cert\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908544 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-ocp-branding-template\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908569 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-trusted-ca-bundle\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908593 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-cliconfig\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908632 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvqph\" (UniqueName: \"kubernetes.io/projected/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-kube-api-access-dvqph\") pod \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\" (UID: \"e1ad5e9b-55f0-441a-84f7-d0f30410abe8\") " Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.908940 4706 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.910256 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.910280 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.910319 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.911396 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.913912 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.914023 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.914616 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.914707 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-kube-api-access-dvqph" (OuterVolumeSpecName: "kube-api-access-dvqph") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "kube-api-access-dvqph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.914815 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.914904 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.915073 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.915572 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:02 crc kubenswrapper[4706]: I1127 07:12:02.915589 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e1ad5e9b-55f0-441a-84f7-d0f30410abe8" (UID: "e1ad5e9b-55f0-441a-84f7-d0f30410abe8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010698 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010741 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010757 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010770 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010784 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010797 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010808 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvqph\" (UniqueName: \"kubernetes.io/projected/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-kube-api-access-dvqph\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010821 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010836 4706 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010850 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010864 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010876 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.010889 4706 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ad5e9b-55f0-441a-84f7-d0f30410abe8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.213309 4706 generic.go:334] "Generic (PLEG): container finished" podID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" containerID="8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e" exitCode=0 Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.213356 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" event={"ID":"e1ad5e9b-55f0-441a-84f7-d0f30410abe8","Type":"ContainerDied","Data":"8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e"} Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.213462 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" event={"ID":"e1ad5e9b-55f0-441a-84f7-d0f30410abe8","Type":"ContainerDied","Data":"a4298ebea9a89eeedc168ff272328fb2566432a4c45cb02c34f4937152297f3c"} Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.213372 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fqglc" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.213616 4706 scope.go:117] "RemoveContainer" containerID="8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.238573 4706 scope.go:117] "RemoveContainer" containerID="8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e" Nov 27 07:12:03 crc kubenswrapper[4706]: E1127 07:12:03.239417 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e\": container with ID starting with 8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e not found: ID does not exist" containerID="8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.239467 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e"} err="failed to get container status \"8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e\": rpc error: code = NotFound desc = could not find container \"8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e\": container with ID starting with 8dfd164f6a437cd2f7e16ed42e34a2eba411ba4b21da58c59b7f6344a7a6ae1e not found: ID does not exist" Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.252665 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fqglc"] Nov 27 07:12:03 crc kubenswrapper[4706]: I1127 07:12:03.256092 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fqglc"] Nov 27 07:12:04 crc kubenswrapper[4706]: I1127 07:12:04.790801 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" path="/var/lib/kubelet/pods/e1ad5e9b-55f0-441a-84f7-d0f30410abe8/volumes" Nov 27 07:12:05 crc kubenswrapper[4706]: I1127 07:12:05.515466 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:12:05 crc kubenswrapper[4706]: I1127 07:12:05.580873 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:12:05 crc kubenswrapper[4706]: I1127 07:12:05.867089 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:12:07 crc kubenswrapper[4706]: I1127 07:12:07.627996 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.144684 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w77l6"] Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.144913 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w77l6" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="registry-server" containerID="cri-o://b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02" gracePeriod=2 Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.458691 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.578671 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc48v\" (UniqueName: \"kubernetes.io/projected/3acd7391-9434-4efe-9526-ba40938af59e-kube-api-access-mc48v\") pod \"3acd7391-9434-4efe-9526-ba40938af59e\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.578734 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-utilities\") pod \"3acd7391-9434-4efe-9526-ba40938af59e\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.578803 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-catalog-content\") pod \"3acd7391-9434-4efe-9526-ba40938af59e\" (UID: \"3acd7391-9434-4efe-9526-ba40938af59e\") " Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.579976 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-utilities" (OuterVolumeSpecName: "utilities") pod "3acd7391-9434-4efe-9526-ba40938af59e" (UID: "3acd7391-9434-4efe-9526-ba40938af59e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.584835 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3acd7391-9434-4efe-9526-ba40938af59e-kube-api-access-mc48v" (OuterVolumeSpecName: "kube-api-access-mc48v") pod "3acd7391-9434-4efe-9526-ba40938af59e" (UID: "3acd7391-9434-4efe-9526-ba40938af59e"). InnerVolumeSpecName "kube-api-access-mc48v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.650354 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3acd7391-9434-4efe-9526-ba40938af59e" (UID: "3acd7391-9434-4efe-9526-ba40938af59e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.679916 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.679964 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc48v\" (UniqueName: \"kubernetes.io/projected/3acd7391-9434-4efe-9526-ba40938af59e-kube-api-access-mc48v\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.679985 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acd7391-9434-4efe-9526-ba40938af59e-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.945620 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5db964fdbd-9vjvs"] Nov 27 07:12:08 crc kubenswrapper[4706]: E1127 07:12:08.946175 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="extract-utilities" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.946265 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="extract-utilities" Nov 27 07:12:08 crc kubenswrapper[4706]: E1127 07:12:08.946326 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="registry-server" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.946390 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="registry-server" Nov 27 07:12:08 crc kubenswrapper[4706]: E1127 07:12:08.946457 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" containerName="oauth-openshift" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.946512 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" containerName="oauth-openshift" Nov 27 07:12:08 crc kubenswrapper[4706]: E1127 07:12:08.946567 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="registry-server" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.946618 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="registry-server" Nov 27 07:12:08 crc kubenswrapper[4706]: E1127 07:12:08.946686 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="extract-content" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.946738 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="extract-content" Nov 27 07:12:08 crc kubenswrapper[4706]: E1127 07:12:08.946795 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="extract-content" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.946846 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="extract-content" Nov 27 07:12:08 crc kubenswrapper[4706]: E1127 07:12:08.946903 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="extract-utilities" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.946957 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="extract-utilities" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.947113 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e1e493-c2ef-42fd-b6ca-41c9bc0cff69" containerName="registry-server" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.947204 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3acd7391-9434-4efe-9526-ba40938af59e" containerName="registry-server" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.947300 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1ad5e9b-55f0-441a-84f7-d0f30410abe8" containerName="oauth-openshift" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.948000 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.950280 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.950488 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.950592 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.951433 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.951635 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.951805 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.952943 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.953176 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.953410 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.953531 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.954656 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.960540 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.961412 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.965306 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5db964fdbd-9vjvs"] Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.966761 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 07:12:08 crc kubenswrapper[4706]: I1127 07:12:08.969188 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.077401 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.083921 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.083983 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-login\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-audit-policies\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084086 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-session\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084115 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084196 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084237 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgngp\" (UniqueName: \"kubernetes.io/projected/2a2d8f23-6038-4782-8432-5dbbac413dec-kube-api-access-fgngp\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084357 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-error\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084431 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084484 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a2d8f23-6038-4782-8432-5dbbac413dec-audit-dir\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084523 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084597 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.084629 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.128892 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185612 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185695 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185737 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185770 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-login\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185806 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185837 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-audit-policies\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-session\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185904 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.185977 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.186011 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgngp\" (UniqueName: \"kubernetes.io/projected/2a2d8f23-6038-4782-8432-5dbbac413dec-kube-api-access-fgngp\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.186061 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-error\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.186110 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.186156 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a2d8f23-6038-4782-8432-5dbbac413dec-audit-dir\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.186195 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.187143 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.187314 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-service-ca\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.187442 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a2d8f23-6038-4782-8432-5dbbac413dec-audit-dir\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.191692 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-error\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.193409 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.193965 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.194422 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-router-certs\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.194456 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.194629 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a2d8f23-6038-4782-8432-5dbbac413dec-audit-policies\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.194774 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.197672 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-user-template-login\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.198427 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.206561 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a2d8f23-6038-4782-8432-5dbbac413dec-v4-0-config-system-session\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.210502 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgngp\" (UniqueName: \"kubernetes.io/projected/2a2d8f23-6038-4782-8432-5dbbac413dec-kube-api-access-fgngp\") pod \"oauth-openshift-5db964fdbd-9vjvs\" (UID: \"2a2d8f23-6038-4782-8432-5dbbac413dec\") " pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.252351 4706 generic.go:334] "Generic (PLEG): container finished" podID="3acd7391-9434-4efe-9526-ba40938af59e" containerID="b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02" exitCode=0 Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.252427 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w77l6" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.252470 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w77l6" event={"ID":"3acd7391-9434-4efe-9526-ba40938af59e","Type":"ContainerDied","Data":"b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02"} Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.253296 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w77l6" event={"ID":"3acd7391-9434-4efe-9526-ba40938af59e","Type":"ContainerDied","Data":"cccf956716a7d176ab33b32ff7bbc33ddf5863f38c054dff54cb3f71016d5d94"} Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.253332 4706 scope.go:117] "RemoveContainer" containerID="b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.275021 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.278736 4706 scope.go:117] "RemoveContainer" containerID="a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.285760 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w77l6"] Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.289376 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w77l6"] Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.308896 4706 scope.go:117] "RemoveContainer" containerID="378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.337250 4706 scope.go:117] "RemoveContainer" containerID="b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02" Nov 27 07:12:09 crc kubenswrapper[4706]: E1127 07:12:09.337771 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02\": container with ID starting with b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02 not found: ID does not exist" containerID="b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.337803 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02"} err="failed to get container status \"b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02\": rpc error: code = NotFound desc = could not find container \"b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02\": container with ID starting with b67bc26cfcd291f2e11287f962ca5d6c91aa40912f857d3d2e55e8f1ad83be02 not found: ID does not exist" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.337824 4706 scope.go:117] "RemoveContainer" containerID="a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76" Nov 27 07:12:09 crc kubenswrapper[4706]: E1127 07:12:09.338073 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76\": container with ID starting with a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76 not found: ID does not exist" containerID="a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.338104 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76"} err="failed to get container status \"a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76\": rpc error: code = NotFound desc = could not find container \"a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76\": container with ID starting with a812326cd7cfaffda1fe1e952102faf3431626888f58200dacbd7eb320bace76 not found: ID does not exist" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.338122 4706 scope.go:117] "RemoveContainer" containerID="378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb" Nov 27 07:12:09 crc kubenswrapper[4706]: E1127 07:12:09.338891 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb\": container with ID starting with 378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb not found: ID does not exist" containerID="378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.338957 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb"} err="failed to get container status \"378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb\": rpc error: code = NotFound desc = could not find container \"378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb\": container with ID starting with 378165609ac2a29911e531abeabb87cd9a4384377979d852e518de3b8d118efb not found: ID does not exist" Nov 27 07:12:09 crc kubenswrapper[4706]: I1127 07:12:09.494745 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5db964fdbd-9vjvs"] Nov 27 07:12:09 crc kubenswrapper[4706]: W1127 07:12:09.502958 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a2d8f23_6038_4782_8432_5dbbac413dec.slice/crio-3513de968fe2261369463ed668d445e85f18762d8f791c19842ba87e834bb0a2 WatchSource:0}: Error finding container 3513de968fe2261369463ed668d445e85f18762d8f791c19842ba87e834bb0a2: Status 404 returned error can't find the container with id 3513de968fe2261369463ed668d445e85f18762d8f791c19842ba87e834bb0a2 Nov 27 07:12:10 crc kubenswrapper[4706]: I1127 07:12:10.260756 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" event={"ID":"2a2d8f23-6038-4782-8432-5dbbac413dec","Type":"ContainerStarted","Data":"25ab962f66b62901216fba2cada0507b289cb422c2411c77efc7170fb79a6cd6"} Nov 27 07:12:10 crc kubenswrapper[4706]: I1127 07:12:10.261103 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" event={"ID":"2a2d8f23-6038-4782-8432-5dbbac413dec","Type":"ContainerStarted","Data":"3513de968fe2261369463ed668d445e85f18762d8f791c19842ba87e834bb0a2"} Nov 27 07:12:10 crc kubenswrapper[4706]: I1127 07:12:10.261149 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:10 crc kubenswrapper[4706]: I1127 07:12:10.265356 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" Nov 27 07:12:10 crc kubenswrapper[4706]: I1127 07:12:10.280884 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5db964fdbd-9vjvs" podStartSLOduration=33.280853852999996 podStartE2EDuration="33.280853853s" podCreationTimestamp="2025-11-27 07:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:12:10.27929504 +0000 UTC m=+214.168885890" watchObservedRunningTime="2025-11-27 07:12:10.280853853 +0000 UTC m=+214.170444703" Nov 27 07:12:10 crc kubenswrapper[4706]: I1127 07:12:10.785999 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3acd7391-9434-4efe-9526-ba40938af59e" path="/var/lib/kubelet/pods/3acd7391-9434-4efe-9526-ba40938af59e/volumes" Nov 27 07:12:11 crc kubenswrapper[4706]: I1127 07:12:11.944143 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tnxnn"] Nov 27 07:12:11 crc kubenswrapper[4706]: I1127 07:12:11.944394 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tnxnn" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="registry-server" containerID="cri-o://297311e6901b449baf19c5846251eade8677cb9b2ea832939bbbf599b3ded2ec" gracePeriod=2 Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.297113 4706 generic.go:334] "Generic (PLEG): container finished" podID="24597f11-1294-4631-82db-4a62bad47e8c" containerID="297311e6901b449baf19c5846251eade8677cb9b2ea832939bbbf599b3ded2ec" exitCode=0 Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.297179 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnxnn" event={"ID":"24597f11-1294-4631-82db-4a62bad47e8c","Type":"ContainerDied","Data":"297311e6901b449baf19c5846251eade8677cb9b2ea832939bbbf599b3ded2ec"} Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.359744 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.531114 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-utilities\") pod \"24597f11-1294-4631-82db-4a62bad47e8c\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.531171 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-catalog-content\") pod \"24597f11-1294-4631-82db-4a62bad47e8c\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.531198 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcpm4\" (UniqueName: \"kubernetes.io/projected/24597f11-1294-4631-82db-4a62bad47e8c-kube-api-access-lcpm4\") pod \"24597f11-1294-4631-82db-4a62bad47e8c\" (UID: \"24597f11-1294-4631-82db-4a62bad47e8c\") " Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.531788 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-utilities" (OuterVolumeSpecName: "utilities") pod "24597f11-1294-4631-82db-4a62bad47e8c" (UID: "24597f11-1294-4631-82db-4a62bad47e8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.536646 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24597f11-1294-4631-82db-4a62bad47e8c-kube-api-access-lcpm4" (OuterVolumeSpecName: "kube-api-access-lcpm4") pod "24597f11-1294-4631-82db-4a62bad47e8c" (UID: "24597f11-1294-4631-82db-4a62bad47e8c"). InnerVolumeSpecName "kube-api-access-lcpm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.613004 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24597f11-1294-4631-82db-4a62bad47e8c" (UID: "24597f11-1294-4631-82db-4a62bad47e8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.632457 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.632504 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24597f11-1294-4631-82db-4a62bad47e8c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:12 crc kubenswrapper[4706]: I1127 07:12:12.632516 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcpm4\" (UniqueName: \"kubernetes.io/projected/24597f11-1294-4631-82db-4a62bad47e8c-kube-api-access-lcpm4\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:13 crc kubenswrapper[4706]: I1127 07:12:13.305789 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnxnn" event={"ID":"24597f11-1294-4631-82db-4a62bad47e8c","Type":"ContainerDied","Data":"7ce56a908c327b0dc78b94207562540b50d7d1db9757c94b4575a7545e4f8e44"} Nov 27 07:12:13 crc kubenswrapper[4706]: I1127 07:12:13.305856 4706 scope.go:117] "RemoveContainer" containerID="297311e6901b449baf19c5846251eade8677cb9b2ea832939bbbf599b3ded2ec" Nov 27 07:12:13 crc kubenswrapper[4706]: I1127 07:12:13.305856 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnxnn" Nov 27 07:12:13 crc kubenswrapper[4706]: I1127 07:12:13.318671 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tnxnn"] Nov 27 07:12:13 crc kubenswrapper[4706]: I1127 07:12:13.323986 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tnxnn"] Nov 27 07:12:13 crc kubenswrapper[4706]: I1127 07:12:13.326542 4706 scope.go:117] "RemoveContainer" containerID="e990c1c678846daca31db721831bf84b68c2e8161115647e9b651e1485484cbb" Nov 27 07:12:13 crc kubenswrapper[4706]: I1127 07:12:13.341562 4706 scope.go:117] "RemoveContainer" containerID="ee9cddecdab2baa44a4990bfdcce59398f66cfc9f06825deb2fcd36ff95acc1f" Nov 27 07:12:14 crc kubenswrapper[4706]: I1127 07:12:14.782779 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24597f11-1294-4631-82db-4a62bad47e8c" path="/var/lib/kubelet/pods/24597f11-1294-4631-82db-4a62bad47e8c/volumes" Nov 27 07:12:15 crc kubenswrapper[4706]: I1127 07:12:15.177988 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:12:15 crc kubenswrapper[4706]: I1127 07:12:15.178037 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:12:15 crc kubenswrapper[4706]: I1127 07:12:15.178076 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:12:15 crc kubenswrapper[4706]: I1127 07:12:15.178476 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:12:15 crc kubenswrapper[4706]: I1127 07:12:15.178525 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f" gracePeriod=600 Nov 27 07:12:15 crc kubenswrapper[4706]: I1127 07:12:15.319248 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f" exitCode=0 Nov 27 07:12:15 crc kubenswrapper[4706]: I1127 07:12:15.319298 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f"} Nov 27 07:12:16 crc kubenswrapper[4706]: I1127 07:12:16.325540 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"8bcee89027fe370848c3797738380643eda24e7a750e5b718af96d44bf36e027"} Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.979481 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.980431 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="extract-utilities" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.980446 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="extract-utilities" Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.980492 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="extract-content" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.980501 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="extract-content" Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.980513 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="registry-server" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.980521 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="registry-server" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.980673 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="24597f11-1294-4631-82db-4a62bad47e8c" containerName="registry-server" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981132 4706 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981159 4706 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981329 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.981344 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981516 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.981564 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981585 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.981626 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981648 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981621 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217" gracePeriod=15 Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.981675 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981693 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981693 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974" gracePeriod=15 Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.981715 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981731 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981755 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f" gracePeriod=15 Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.981761 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981779 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 07:12:29 crc kubenswrapper[4706]: E1127 07:12:29.981797 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981697 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c" gracePeriod=15 Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981815 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.981777 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744" gracePeriod=15 Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.982129 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.982157 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.982181 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.982207 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.982262 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.982290 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 07:12:29 crc kubenswrapper[4706]: I1127 07:12:29.986799 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 27 07:12:30 crc kubenswrapper[4706]: E1127 07:12:30.051330 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157091 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157164 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157211 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157293 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157311 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157330 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157360 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.157538 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258650 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258705 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258749 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258780 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258803 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258803 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258830 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258846 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258866 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258860 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258822 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.258953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.259017 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.259092 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.259112 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.352832 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:30 crc kubenswrapper[4706]: W1127 07:12:30.374336 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-6b7f4cef52ca2ac139f27b9e3d091db2e0eaac261a43f4477ac90d2888beb6c2 WatchSource:0}: Error finding container 6b7f4cef52ca2ac139f27b9e3d091db2e0eaac261a43f4477ac90d2888beb6c2: Status 404 returned error can't find the container with id 6b7f4cef52ca2ac139f27b9e3d091db2e0eaac261a43f4477ac90d2888beb6c2 Nov 27 07:12:30 crc kubenswrapper[4706]: E1127 07:12:30.377447 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.153:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187bcb9c1b2c15e3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 07:12:30.376891875 +0000 UTC m=+234.266482685,LastTimestamp:2025-11-27 07:12:30.376891875 +0000 UTC m=+234.266482685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.415817 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.417251 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.418103 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c" exitCode=0 Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.418245 4706 scope.go:117] "RemoveContainer" containerID="8a7f53a57f4bee7433b2fe80fdf77b248c6140d1876635244a1de5a58a657116" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.418275 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f" exitCode=0 Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.418307 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974" exitCode=0 Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.418329 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744" exitCode=2 Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.420362 4706 generic.go:334] "Generic (PLEG): container finished" podID="e14e4ff4-0455-402f-8299-044856ac016d" containerID="9d76ea3d0e4555adfa880c6e49bb9a6506b75258ca9541ed2a1fc7a1b1d6065b" exitCode=0 Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.420419 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e14e4ff4-0455-402f-8299-044856ac016d","Type":"ContainerDied","Data":"9d76ea3d0e4555adfa880c6e49bb9a6506b75258ca9541ed2a1fc7a1b1d6065b"} Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.420889 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:30 crc kubenswrapper[4706]: I1127 07:12:30.421892 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6b7f4cef52ca2ac139f27b9e3d091db2e0eaac261a43f4477ac90d2888beb6c2"} Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.433358 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.439590 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa"} Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.440698 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:31 crc kubenswrapper[4706]: E1127 07:12:31.441898 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.697589 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.698289 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.779307 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e14e4ff4-0455-402f-8299-044856ac016d-kube-api-access\") pod \"e14e4ff4-0455-402f-8299-044856ac016d\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.779430 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-var-lock\") pod \"e14e4ff4-0455-402f-8299-044856ac016d\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.779504 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-kubelet-dir\") pod \"e14e4ff4-0455-402f-8299-044856ac016d\" (UID: \"e14e4ff4-0455-402f-8299-044856ac016d\") " Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.779565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-var-lock" (OuterVolumeSpecName: "var-lock") pod "e14e4ff4-0455-402f-8299-044856ac016d" (UID: "e14e4ff4-0455-402f-8299-044856ac016d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.779673 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e14e4ff4-0455-402f-8299-044856ac016d" (UID: "e14e4ff4-0455-402f-8299-044856ac016d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.779876 4706 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.779899 4706 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e14e4ff4-0455-402f-8299-044856ac016d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.784654 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e14e4ff4-0455-402f-8299-044856ac016d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e14e4ff4-0455-402f-8299-044856ac016d" (UID: "e14e4ff4-0455-402f-8299-044856ac016d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:12:31 crc kubenswrapper[4706]: I1127 07:12:31.881210 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e14e4ff4-0455-402f-8299-044856ac016d-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.351773 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.353406 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.354167 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.354696 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.447734 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e14e4ff4-0455-402f-8299-044856ac016d","Type":"ContainerDied","Data":"dfc6f07343f40b0868a81db13132310d3348e87f1223c03d829060370a677c5f"} Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.447772 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfc6f07343f40b0868a81db13132310d3348e87f1223c03d829060370a677c5f" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.447785 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.450536 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.451358 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217" exitCode=0 Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.451432 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.451474 4706 scope.go:117] "RemoveContainer" containerID="e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c" Nov 27 07:12:32 crc kubenswrapper[4706]: E1127 07:12:32.452199 4706 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.474351 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.475019 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.477534 4706 scope.go:117] "RemoveContainer" containerID="90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.489594 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.489647 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.489775 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.490003 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.490037 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.490055 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.498838 4706 scope.go:117] "RemoveContainer" containerID="b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.519530 4706 scope.go:117] "RemoveContainer" containerID="3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.532972 4706 scope.go:117] "RemoveContainer" containerID="44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.553418 4706 scope.go:117] "RemoveContainer" containerID="fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.578648 4706 scope.go:117] "RemoveContainer" containerID="e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c" Nov 27 07:12:32 crc kubenswrapper[4706]: E1127 07:12:32.579192 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\": container with ID starting with e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c not found: ID does not exist" containerID="e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.579240 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c"} err="failed to get container status \"e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\": rpc error: code = NotFound desc = could not find container \"e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c\": container with ID starting with e8071a8941cf2c87a4bf455eb283e60f9e25026591b5fb6548316e8aa85e951c not found: ID does not exist" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.579266 4706 scope.go:117] "RemoveContainer" containerID="90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f" Nov 27 07:12:32 crc kubenswrapper[4706]: E1127 07:12:32.579813 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\": container with ID starting with 90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f not found: ID does not exist" containerID="90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.579840 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f"} err="failed to get container status \"90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\": rpc error: code = NotFound desc = could not find container \"90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f\": container with ID starting with 90235ce2b3e6c328dddb2923d6ae7e8efd4a267ccd147f45a4e641eef778100f not found: ID does not exist" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.579859 4706 scope.go:117] "RemoveContainer" containerID="b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974" Nov 27 07:12:32 crc kubenswrapper[4706]: E1127 07:12:32.580343 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\": container with ID starting with b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974 not found: ID does not exist" containerID="b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.580373 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974"} err="failed to get container status \"b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\": rpc error: code = NotFound desc = could not find container \"b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974\": container with ID starting with b45407009bbe973a51a19a34f7866819d1ef646e3e4f38b2e074ca65f91ed974 not found: ID does not exist" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.580393 4706 scope.go:117] "RemoveContainer" containerID="3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744" Nov 27 07:12:32 crc kubenswrapper[4706]: E1127 07:12:32.580884 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\": container with ID starting with 3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744 not found: ID does not exist" containerID="3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.580912 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744"} err="failed to get container status \"3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\": rpc error: code = NotFound desc = could not find container \"3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744\": container with ID starting with 3b967314d1bdb18ff2770885e3863a6f776d0b11bc161302b8bf8b09a8bf4744 not found: ID does not exist" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.580929 4706 scope.go:117] "RemoveContainer" containerID="44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217" Nov 27 07:12:32 crc kubenswrapper[4706]: E1127 07:12:32.581398 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\": container with ID starting with 44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217 not found: ID does not exist" containerID="44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.581426 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217"} err="failed to get container status \"44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\": rpc error: code = NotFound desc = could not find container \"44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217\": container with ID starting with 44ce57df2b099e2c7cb85564bb9d4402951415f910045c7db76ed8c8f1805217 not found: ID does not exist" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.581445 4706 scope.go:117] "RemoveContainer" containerID="fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159" Nov 27 07:12:32 crc kubenswrapper[4706]: E1127 07:12:32.581901 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\": container with ID starting with fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159 not found: ID does not exist" containerID="fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.581930 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159"} err="failed to get container status \"fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\": rpc error: code = NotFound desc = could not find container \"fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159\": container with ID starting with fa99dd453d032217151ebb08922fe32150bb5e785f2a3b126cc18f1f935ce159 not found: ID does not exist" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.590853 4706 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.590888 4706 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.590899 4706 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.766871 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.767543 4706 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:32 crc kubenswrapper[4706]: I1127 07:12:32.786029 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 27 07:12:36 crc kubenswrapper[4706]: E1127 07:12:36.593045 4706 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.153:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187bcb9c1b2c15e3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 07:12:30.376891875 +0000 UTC m=+234.266482685,LastTimestamp:2025-11-27 07:12:30.376891875 +0000 UTC m=+234.266482685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 07:12:36 crc kubenswrapper[4706]: I1127 07:12:36.779948 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:37 crc kubenswrapper[4706]: E1127 07:12:37.540917 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:37 crc kubenswrapper[4706]: E1127 07:12:37.541629 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:37 crc kubenswrapper[4706]: E1127 07:12:37.542286 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:37 crc kubenswrapper[4706]: E1127 07:12:37.542723 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:37 crc kubenswrapper[4706]: E1127 07:12:37.543180 4706 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:37 crc kubenswrapper[4706]: I1127 07:12:37.543283 4706 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 27 07:12:37 crc kubenswrapper[4706]: E1127 07:12:37.543770 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="200ms" Nov 27 07:12:37 crc kubenswrapper[4706]: E1127 07:12:37.744941 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="400ms" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.146987 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="800ms" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.338642 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:12:38Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:12:38Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:12:38Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T07:12:38Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.339499 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.340005 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.340600 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.341049 4706 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.341082 4706 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 07:12:38 crc kubenswrapper[4706]: E1127 07:12:38.947687 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="1.6s" Nov 27 07:12:40 crc kubenswrapper[4706]: E1127 07:12:40.548863 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="3.2s" Nov 27 07:12:40 crc kubenswrapper[4706]: E1127 07:12:40.819880 4706 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" volumeName="registry-storage" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.518558 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.518865 4706 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de" exitCode=1 Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.518906 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de"} Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.519438 4706 scope.go:117] "RemoveContainer" containerID="8bc85af06afe93a8b1fe480f25b0cd48ed459d4440972d9f7baeec74236fb2de" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.520400 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.520648 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:43 crc kubenswrapper[4706]: E1127 07:12:43.750028 4706 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.153:6443: connect: connection refused" interval="6.4s" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.776477 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.777328 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.777562 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.795425 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.795461 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:43 crc kubenswrapper[4706]: E1127 07:12:43.795834 4706 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:43 crc kubenswrapper[4706]: I1127 07:12:43.796334 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:43 crc kubenswrapper[4706]: W1127 07:12:43.818669 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-26b6cbfba0e79795625ed63a9e58924f92ea0c89e6519906c2aeb47f6efbc34b WatchSource:0}: Error finding container 26b6cbfba0e79795625ed63a9e58924f92ea0c89e6519906c2aeb47f6efbc34b: Status 404 returned error can't find the container with id 26b6cbfba0e79795625ed63a9e58924f92ea0c89e6519906c2aeb47f6efbc34b Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.527696 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.528004 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"106ef90d3f41751053b071e5035ab2feae538908e9dcaa2af08b36333a87f5bb"} Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.528885 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.529370 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.529782 4706 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="65c53348d9b93b5a0cf3c1a98c5c95d38d42424b5eb1cdef01d0aa26389d259e" exitCode=0 Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.529854 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"65c53348d9b93b5a0cf3c1a98c5c95d38d42424b5eb1cdef01d0aa26389d259e"} Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.529904 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"26b6cbfba0e79795625ed63a9e58924f92ea0c89e6519906c2aeb47f6efbc34b"} Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.530163 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.530188 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:44 crc kubenswrapper[4706]: E1127 07:12:44.530473 4706 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.530516 4706 status_manager.go:851] "Failed to get status for pod" podUID="e14e4ff4-0455-402f-8299-044856ac016d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:44 crc kubenswrapper[4706]: I1127 07:12:44.530744 4706 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.153:6443: connect: connection refused" Nov 27 07:12:45 crc kubenswrapper[4706]: I1127 07:12:45.536508 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a0db6d5c3dfd9edfec9bbced70d4405acdc951b1ec4727c6c97edba7e4afbac5"} Nov 27 07:12:45 crc kubenswrapper[4706]: I1127 07:12:45.536807 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"379125079cff210f211e7114198d30ea5ad8d85281288496170432e55e5b198b"} Nov 27 07:12:45 crc kubenswrapper[4706]: I1127 07:12:45.536817 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1f489eead9e347c6242cede5149ea0be618da7ce47b6401725aef891d3d5aca5"} Nov 27 07:12:45 crc kubenswrapper[4706]: I1127 07:12:45.536825 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"36c5814b29bf769a7bc14bde3f51b8d5256f38af586f59c0341bf64b12f324c7"} Nov 27 07:12:46 crc kubenswrapper[4706]: I1127 07:12:46.547867 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"65c2096ca65e53f94eb98ca459dffb823f11fbd00b5cd4c63b0796fff5cbd7e1"} Nov 27 07:12:46 crc kubenswrapper[4706]: I1127 07:12:46.548110 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:46 crc kubenswrapper[4706]: I1127 07:12:46.548123 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:46 crc kubenswrapper[4706]: I1127 07:12:46.548280 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:47 crc kubenswrapper[4706]: I1127 07:12:47.107618 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:12:47 crc kubenswrapper[4706]: I1127 07:12:47.111124 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:12:47 crc kubenswrapper[4706]: I1127 07:12:47.552924 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:12:48 crc kubenswrapper[4706]: I1127 07:12:48.797090 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:48 crc kubenswrapper[4706]: I1127 07:12:48.797166 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:48 crc kubenswrapper[4706]: I1127 07:12:48.806768 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:51 crc kubenswrapper[4706]: I1127 07:12:51.559608 4706 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:51 crc kubenswrapper[4706]: I1127 07:12:51.617418 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8f7fd49e-c7b7-4a7a-ad1b-5994f97b0c44" Nov 27 07:12:52 crc kubenswrapper[4706]: I1127 07:12:52.581298 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:52 crc kubenswrapper[4706]: I1127 07:12:52.581333 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:52 crc kubenswrapper[4706]: I1127 07:12:52.584483 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8f7fd49e-c7b7-4a7a-ad1b-5994f97b0c44" Nov 27 07:12:52 crc kubenswrapper[4706]: I1127 07:12:52.586063 4706 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://36c5814b29bf769a7bc14bde3f51b8d5256f38af586f59c0341bf64b12f324c7" Nov 27 07:12:52 crc kubenswrapper[4706]: I1127 07:12:52.586099 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:12:53 crc kubenswrapper[4706]: I1127 07:12:53.585387 4706 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:53 crc kubenswrapper[4706]: I1127 07:12:53.585417 4706 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ee814c58-bb88-4206-bf64-22abdbdeae78" Nov 27 07:12:53 crc kubenswrapper[4706]: I1127 07:12:53.588838 4706 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8f7fd49e-c7b7-4a7a-ad1b-5994f97b0c44" Nov 27 07:12:58 crc kubenswrapper[4706]: I1127 07:12:58.455080 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 07:13:01 crc kubenswrapper[4706]: I1127 07:13:01.126066 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 07:13:01 crc kubenswrapper[4706]: I1127 07:13:01.201186 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 07:13:01 crc kubenswrapper[4706]: I1127 07:13:01.632291 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 07:13:01 crc kubenswrapper[4706]: I1127 07:13:01.638262 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 07:13:01 crc kubenswrapper[4706]: I1127 07:13:01.696271 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 07:13:01 crc kubenswrapper[4706]: I1127 07:13:01.847285 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 07:13:01 crc kubenswrapper[4706]: I1127 07:13:01.910054 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.133892 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.229360 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.233784 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.255043 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.262342 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.388518 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.427471 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.537161 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.545027 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.896173 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 07:13:02 crc kubenswrapper[4706]: I1127 07:13:02.917843 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 07:13:03 crc kubenswrapper[4706]: I1127 07:13:03.175728 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 07:13:03 crc kubenswrapper[4706]: I1127 07:13:03.322469 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 07:13:03 crc kubenswrapper[4706]: I1127 07:13:03.335061 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 07:13:03 crc kubenswrapper[4706]: I1127 07:13:03.730555 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 07:13:03 crc kubenswrapper[4706]: I1127 07:13:03.738858 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 07:13:03 crc kubenswrapper[4706]: I1127 07:13:03.754573 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 07:13:03 crc kubenswrapper[4706]: I1127 07:13:03.790919 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.021953 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.034810 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.084442 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.162672 4706 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.307601 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.438426 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.511382 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.517442 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.603887 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.651895 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.678359 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.773676 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.777630 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.860137 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.888081 4706 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.949428 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 07:13:04 crc kubenswrapper[4706]: I1127 07:13:04.950666 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.038961 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.072876 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.112937 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.127529 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.173018 4706 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.244150 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.254888 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.370516 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.543415 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.631992 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.656777 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.676867 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.812150 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.824105 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 07:13:05 crc kubenswrapper[4706]: I1127 07:13:05.873825 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.026959 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.042546 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.205940 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.286164 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.289948 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.322960 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.415044 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.431098 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.526639 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.659292 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.747394 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.871397 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 07:13:06 crc kubenswrapper[4706]: I1127 07:13:06.915742 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.066500 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.069140 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.104756 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.119706 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.345532 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.377785 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.403397 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.438126 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.461197 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.616242 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.667869 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.688252 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.704179 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.722203 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.760793 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.852034 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.852393 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.866814 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.928496 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 07:13:07 crc kubenswrapper[4706]: I1127 07:13:07.971131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.075324 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.125369 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.166909 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.172396 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.177138 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.236477 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.416358 4706 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.439429 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.620492 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.841115 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.943751 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 07:13:08 crc kubenswrapper[4706]: I1127 07:13:08.995975 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.009940 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.031094 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.034862 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.044721 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.053907 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.072628 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.161373 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.163552 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.201490 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.408807 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.426131 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.464980 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.500810 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.633481 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.726140 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.807811 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.807948 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.816940 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.866735 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.922646 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 07:13:09 crc kubenswrapper[4706]: I1127 07:13:09.944320 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.068400 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.095992 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.156980 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.319636 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.554793 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.573621 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.640811 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.657790 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.666792 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.668585 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.725974 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.737619 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.788693 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.822174 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.843876 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.899293 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 07:13:10 crc kubenswrapper[4706]: I1127 07:13:10.940072 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.033031 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.050288 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.151897 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.189246 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.206470 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.276583 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.315384 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.345061 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.411605 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.434677 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.497469 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.529287 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.562898 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.685759 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.701243 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.711942 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.739452 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.767671 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.845750 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.854182 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.864867 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.888766 4706 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.901273 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.911215 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 07:13:11 crc kubenswrapper[4706]: I1127 07:13:11.955696 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.053782 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.082799 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.238123 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.251552 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.256491 4706 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.260972 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.261022 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.265049 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.275079 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.296088 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.296069682 podStartE2EDuration="21.296069682s" podCreationTimestamp="2025-11-27 07:12:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:13:12.278356335 +0000 UTC m=+276.167947155" watchObservedRunningTime="2025-11-27 07:13:12.296069682 +0000 UTC m=+276.185660502" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.386969 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.413505 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.423368 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.426271 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.443631 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.460096 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.626414 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.719833 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.719842 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.731493 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.828024 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.839273 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.899136 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.905029 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.908338 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.929996 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 07:13:12 crc kubenswrapper[4706]: I1127 07:13:12.972178 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.101201 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.106488 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.308545 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.330237 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.376080 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.425327 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.436355 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.457367 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.476845 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.498634 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.517453 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.598045 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.628922 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.771895 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.779893 4706 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.780211 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa" gracePeriod=5 Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.795962 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 07:13:13 crc kubenswrapper[4706]: I1127 07:13:13.898951 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.021972 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.131771 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.166419 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.271460 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.401515 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.563033 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.566593 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.644685 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.668797 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.676667 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.750126 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.766508 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.789099 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.885983 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.948114 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 07:13:14 crc kubenswrapper[4706]: I1127 07:13:14.983507 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.047030 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.048927 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.138459 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.223766 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.517105 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.544391 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.553759 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.585860 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.587207 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.647689 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.658827 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.713110 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.946933 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 07:13:15 crc kubenswrapper[4706]: I1127 07:13:15.996321 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.020450 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.033216 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.094931 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.211060 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.287436 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.307509 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.458849 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.564777 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.565008 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.666343 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 07:13:16 crc kubenswrapper[4706]: I1127 07:13:16.667698 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 07:13:17 crc kubenswrapper[4706]: I1127 07:13:17.255000 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 07:13:17 crc kubenswrapper[4706]: I1127 07:13:17.276728 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 07:13:17 crc kubenswrapper[4706]: I1127 07:13:17.500414 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.199329 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj2k9"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.200755 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jj2k9" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="registry-server" containerID="cri-o://0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65" gracePeriod=30 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.217644 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tq6jq"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.218202 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tq6jq" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="registry-server" containerID="cri-o://7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099" gracePeriod=30 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.228934 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jgpht"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.233661 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerName="marketplace-operator" containerID="cri-o://11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65" gracePeriod=30 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.234536 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxm5z"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.234764 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gxm5z" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="registry-server" containerID="cri-o://35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52" gracePeriod=30 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.241483 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8vv7v"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.250879 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8vv7v" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="registry-server" containerID="cri-o://a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" gracePeriod=30 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.258831 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fzsrg"] Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.259046 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14e4ff4-0455-402f-8299-044856ac016d" containerName="installer" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.259062 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14e4ff4-0455-402f-8299-044856ac016d" containerName="installer" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.259070 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.259078 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.259157 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.259171 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e14e4ff4-0455-402f-8299-044856ac016d" containerName="installer" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.259557 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.268906 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fzsrg"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.276360 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.279105 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7353c5-075d-4f1d-b18e-2da68cc06652-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.279160 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5zdl\" (UniqueName: \"kubernetes.io/projected/bd7353c5-075d-4f1d-b18e-2da68cc06652-kube-api-access-d5zdl\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.279209 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bd7353c5-075d-4f1d-b18e-2da68cc06652-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.379909 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7353c5-075d-4f1d-b18e-2da68cc06652-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.380242 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5zdl\" (UniqueName: \"kubernetes.io/projected/bd7353c5-075d-4f1d-b18e-2da68cc06652-kube-api-access-d5zdl\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.380279 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bd7353c5-075d-4f1d-b18e-2da68cc06652-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.381428 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd7353c5-075d-4f1d-b18e-2da68cc06652-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.402090 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bd7353c5-075d-4f1d-b18e-2da68cc06652-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.407436 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5zdl\" (UniqueName: \"kubernetes.io/projected/bd7353c5-075d-4f1d-b18e-2da68cc06652-kube-api-access-d5zdl\") pod \"marketplace-operator-79b997595-fzsrg\" (UID: \"bd7353c5-075d-4f1d-b18e-2da68cc06652\") " pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.611802 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f is running failed: container process not found" containerID="a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.612040 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f is running failed: container process not found" containerID="a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.612253 4706 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f is running failed: container process not found" containerID="a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.612287 4706 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-8vv7v" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="registry-server" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.616970 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.622432 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.628735 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.682028 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.682744 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca\") pod \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.682795 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57pcz\" (UniqueName: \"kubernetes.io/projected/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-kube-api-access-57pcz\") pod \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.682850 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-utilities\") pod \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.682908 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-catalog-content\") pod \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.682937 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wcqm\" (UniqueName: \"kubernetes.io/projected/a8ac4a84-ddce-41e8-9da0-efd60117aa67-kube-api-access-8wcqm\") pod \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\" (UID: \"a8ac4a84-ddce-41e8-9da0-efd60117aa67\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.682955 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics\") pod \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\" (UID: \"e22ae6d0-dbdf-4b69-84f0-0643266a7dff\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.683969 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "e22ae6d0-dbdf-4b69-84f0-0643266a7dff" (UID: "e22ae6d0-dbdf-4b69-84f0-0643266a7dff"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.684536 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-utilities" (OuterVolumeSpecName: "utilities") pod "a8ac4a84-ddce-41e8-9da0-efd60117aa67" (UID: "a8ac4a84-ddce-41e8-9da0-efd60117aa67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.686507 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "e22ae6d0-dbdf-4b69-84f0-0643266a7dff" (UID: "e22ae6d0-dbdf-4b69-84f0-0643266a7dff"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.686665 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-kube-api-access-57pcz" (OuterVolumeSpecName: "kube-api-access-57pcz") pod "e22ae6d0-dbdf-4b69-84f0-0643266a7dff" (UID: "e22ae6d0-dbdf-4b69-84f0-0643266a7dff"). InnerVolumeSpecName "kube-api-access-57pcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.688582 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.689204 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8ac4a84-ddce-41e8-9da0-efd60117aa67-kube-api-access-8wcqm" (OuterVolumeSpecName: "kube-api-access-8wcqm") pod "a8ac4a84-ddce-41e8-9da0-efd60117aa67" (UID: "a8ac4a84-ddce-41e8-9da0-efd60117aa67"). InnerVolumeSpecName "kube-api-access-8wcqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.729136 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.738604 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerID="35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52" exitCode=0 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.738659 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxm5z" event={"ID":"3b1176cb-10d1-4516-9ece-d51bb6ca515d","Type":"ContainerDied","Data":"35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.738715 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxm5z" event={"ID":"3b1176cb-10d1-4516-9ece-d51bb6ca515d","Type":"ContainerDied","Data":"ce4ba0714eecf1b51e998693d3716fa7a26cf3aea07ff358c5828944aaf3f325"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.738683 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxm5z" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.738737 4706 scope.go:117] "RemoveContainer" containerID="35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.741342 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8ac4a84-ddce-41e8-9da0-efd60117aa67" (UID: "a8ac4a84-ddce-41e8-9da0-efd60117aa67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.743094 4706 generic.go:334] "Generic (PLEG): container finished" podID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerID="a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" exitCode=0 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.743150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8vv7v" event={"ID":"adf72c72-f7cf-4ed2-9356-ef2c880d8164","Type":"ContainerDied","Data":"a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.743204 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8vv7v" event={"ID":"adf72c72-f7cf-4ed2-9356-ef2c880d8164","Type":"ContainerDied","Data":"42aac8c7c45033838425b5187b8e88e37a15dfee873c45ae2003256a6f2686c3"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.743274 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8vv7v" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.746667 4706 generic.go:334] "Generic (PLEG): container finished" podID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerID="11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65" exitCode=0 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.746742 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" event={"ID":"e22ae6d0-dbdf-4b69-84f0-0643266a7dff","Type":"ContainerDied","Data":"11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.746775 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" event={"ID":"e22ae6d0-dbdf-4b69-84f0-0643266a7dff","Type":"ContainerDied","Data":"e95a9df2c02b5eadf2fd9615aac85af28a55f9771cb7fe3ffbdcfd8198b1ca70"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.746842 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jgpht" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.753669 4706 generic.go:334] "Generic (PLEG): container finished" podID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerID="7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099" exitCode=0 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.753732 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6jq" event={"ID":"43c8fce5-14bd-4d14-b488-50600b596cb5","Type":"ContainerDied","Data":"7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.753759 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tq6jq" event={"ID":"43c8fce5-14bd-4d14-b488-50600b596cb5","Type":"ContainerDied","Data":"3b323d55f9be44626d90defc91a7c25f02545e59c9ab2f0c376d1828cd6851c8"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.753811 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tq6jq" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.759885 4706 generic.go:334] "Generic (PLEG): container finished" podID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerID="0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65" exitCode=0 Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.759919 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj2k9" event={"ID":"a8ac4a84-ddce-41e8-9da0-efd60117aa67","Type":"ContainerDied","Data":"0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.759939 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj2k9" event={"ID":"a8ac4a84-ddce-41e8-9da0-efd60117aa67","Type":"ContainerDied","Data":"bfd342ab8d81c3d8368b0aecfcc292fb1e535870f2d51f160fa14a208a0488d4"} Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.759947 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj2k9" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.766410 4706 scope.go:117] "RemoveContainer" containerID="53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786614 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-utilities\") pod \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786678 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-catalog-content\") pod \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786736 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-utilities\") pod \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786779 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5d64\" (UniqueName: \"kubernetes.io/projected/43c8fce5-14bd-4d14-b488-50600b596cb5-kube-api-access-l5d64\") pod \"43c8fce5-14bd-4d14-b488-50600b596cb5\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786827 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2llrs\" (UniqueName: \"kubernetes.io/projected/adf72c72-f7cf-4ed2-9356-ef2c880d8164-kube-api-access-2llrs\") pod \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\" (UID: \"adf72c72-f7cf-4ed2-9356-ef2c880d8164\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786849 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-catalog-content\") pod \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786866 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j74fc\" (UniqueName: \"kubernetes.io/projected/3b1176cb-10d1-4516-9ece-d51bb6ca515d-kube-api-access-j74fc\") pod \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\" (UID: \"3b1176cb-10d1-4516-9ece-d51bb6ca515d\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786893 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-catalog-content\") pod \"43c8fce5-14bd-4d14-b488-50600b596cb5\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.786925 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-utilities\") pod \"43c8fce5-14bd-4d14-b488-50600b596cb5\" (UID: \"43c8fce5-14bd-4d14-b488-50600b596cb5\") " Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.787173 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.787191 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57pcz\" (UniqueName: \"kubernetes.io/projected/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-kube-api-access-57pcz\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.787205 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.787230 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8ac4a84-ddce-41e8-9da0-efd60117aa67-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.787243 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wcqm\" (UniqueName: \"kubernetes.io/projected/a8ac4a84-ddce-41e8-9da0-efd60117aa67-kube-api-access-8wcqm\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.787255 4706 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e22ae6d0-dbdf-4b69-84f0-0643266a7dff-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.788039 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-utilities" (OuterVolumeSpecName: "utilities") pod "43c8fce5-14bd-4d14-b488-50600b596cb5" (UID: "43c8fce5-14bd-4d14-b488-50600b596cb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.790621 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-utilities" (OuterVolumeSpecName: "utilities") pod "3b1176cb-10d1-4516-9ece-d51bb6ca515d" (UID: "3b1176cb-10d1-4516-9ece-d51bb6ca515d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.797605 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b1176cb-10d1-4516-9ece-d51bb6ca515d-kube-api-access-j74fc" (OuterVolumeSpecName: "kube-api-access-j74fc") pod "3b1176cb-10d1-4516-9ece-d51bb6ca515d" (UID: "3b1176cb-10d1-4516-9ece-d51bb6ca515d"). InnerVolumeSpecName "kube-api-access-j74fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.801066 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-utilities" (OuterVolumeSpecName: "utilities") pod "adf72c72-f7cf-4ed2-9356-ef2c880d8164" (UID: "adf72c72-f7cf-4ed2-9356-ef2c880d8164"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.808670 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jgpht"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.809782 4706 scope.go:117] "RemoveContainer" containerID="2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.809808 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c8fce5-14bd-4d14-b488-50600b596cb5-kube-api-access-l5d64" (OuterVolumeSpecName: "kube-api-access-l5d64") pod "43c8fce5-14bd-4d14-b488-50600b596cb5" (UID: "43c8fce5-14bd-4d14-b488-50600b596cb5"). InnerVolumeSpecName "kube-api-access-l5d64". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.809927 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf72c72-f7cf-4ed2-9356-ef2c880d8164-kube-api-access-2llrs" (OuterVolumeSpecName: "kube-api-access-2llrs") pod "adf72c72-f7cf-4ed2-9356-ef2c880d8164" (UID: "adf72c72-f7cf-4ed2-9356-ef2c880d8164"). InnerVolumeSpecName "kube-api-access-2llrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.810160 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jgpht"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.820698 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj2k9"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.822292 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b1176cb-10d1-4516-9ece-d51bb6ca515d" (UID: "3b1176cb-10d1-4516-9ece-d51bb6ca515d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.827304 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jj2k9"] Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.838680 4706 scope.go:117] "RemoveContainer" containerID="35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.839870 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52\": container with ID starting with 35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52 not found: ID does not exist" containerID="35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.839913 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52"} err="failed to get container status \"35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52\": rpc error: code = NotFound desc = could not find container \"35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52\": container with ID starting with 35796cfb483277871b51311eec44e2402a9a52b60f33c61822ed99a631af8f52 not found: ID does not exist" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.839984 4706 scope.go:117] "RemoveContainer" containerID="53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.840483 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab\": container with ID starting with 53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab not found: ID does not exist" containerID="53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.840547 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab"} err="failed to get container status \"53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab\": rpc error: code = NotFound desc = could not find container \"53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab\": container with ID starting with 53d5143bdc3bfa42270b8b600d5752c33cb15be7fc9576e6b740ce9d01edb2ab not found: ID does not exist" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.840576 4706 scope.go:117] "RemoveContainer" containerID="2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.840854 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d\": container with ID starting with 2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d not found: ID does not exist" containerID="2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.840885 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d"} err="failed to get container status \"2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d\": rpc error: code = NotFound desc = could not find container \"2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d\": container with ID starting with 2f809c479d8daf5808e12eb8296b886fa770bb801cef49d480c9e7e56039874d not found: ID does not exist" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.840906 4706 scope.go:117] "RemoveContainer" containerID="a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.853614 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.870902 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43c8fce5-14bd-4d14-b488-50600b596cb5" (UID: "43c8fce5-14bd-4d14-b488-50600b596cb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888347 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5d64\" (UniqueName: \"kubernetes.io/projected/43c8fce5-14bd-4d14-b488-50600b596cb5-kube-api-access-l5d64\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888385 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2llrs\" (UniqueName: \"kubernetes.io/projected/adf72c72-f7cf-4ed2-9356-ef2c880d8164-kube-api-access-2llrs\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888396 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888404 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j74fc\" (UniqueName: \"kubernetes.io/projected/3b1176cb-10d1-4516-9ece-d51bb6ca515d-kube-api-access-j74fc\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888413 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888422 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c8fce5-14bd-4d14-b488-50600b596cb5-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888431 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1176cb-10d1-4516-9ece-d51bb6ca515d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.888439 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.897063 4706 scope.go:117] "RemoveContainer" containerID="9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.899299 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.909884 4706 scope.go:117] "RemoveContainer" containerID="28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.912325 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adf72c72-f7cf-4ed2-9356-ef2c880d8164" (UID: "adf72c72-f7cf-4ed2-9356-ef2c880d8164"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.947336 4706 scope.go:117] "RemoveContainer" containerID="a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.948059 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f\": container with ID starting with a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f not found: ID does not exist" containerID="a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.948094 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f"} err="failed to get container status \"a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f\": rpc error: code = NotFound desc = could not find container \"a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f\": container with ID starting with a5340ae19c7d6fc6ac6d2da2ae5f01c8a7c9f182264e2ab0adc97164cb861c9f not found: ID does not exist" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.948119 4706 scope.go:117] "RemoveContainer" containerID="9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.949118 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c\": container with ID starting with 9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c not found: ID does not exist" containerID="9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.949195 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c"} err="failed to get container status \"9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c\": rpc error: code = NotFound desc = could not find container \"9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c\": container with ID starting with 9b5dc4d011cd040d346c13de3c968fbd41f07dd276698d19bb39749cc07a3b1c not found: ID does not exist" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.949264 4706 scope.go:117] "RemoveContainer" containerID="28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.949841 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e\": container with ID starting with 28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e not found: ID does not exist" containerID="28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.949894 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e"} err="failed to get container status \"28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e\": rpc error: code = NotFound desc = could not find container \"28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e\": container with ID starting with 28529ff733c7cc4ed3f6d135f68fab5a73de32337c69234a51575091f99aab5e not found: ID does not exist" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.949930 4706 scope.go:117] "RemoveContainer" containerID="11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.954432 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.961989 4706 scope.go:117] "RemoveContainer" containerID="11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65" Nov 27 07:13:18 crc kubenswrapper[4706]: E1127 07:13:18.962519 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65\": container with ID starting with 11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65 not found: ID does not exist" containerID="11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.962548 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65"} err="failed to get container status \"11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65\": rpc error: code = NotFound desc = could not find container \"11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65\": container with ID starting with 11c3f0f479d03ed7fc0cc7bbef8eaeb65a84abe69b197db03dc9f9665e256b65 not found: ID does not exist" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.962568 4706 scope.go:117] "RemoveContainer" containerID="7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.977397 4706 scope.go:117] "RemoveContainer" containerID="0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.989151 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf72c72-f7cf-4ed2-9356-ef2c880d8164-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:18 crc kubenswrapper[4706]: I1127 07:13:18.990245 4706 scope.go:117] "RemoveContainer" containerID="00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.002134 4706 scope.go:117] "RemoveContainer" containerID="7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099" Nov 27 07:13:19 crc kubenswrapper[4706]: E1127 07:13:19.002497 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099\": container with ID starting with 7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099 not found: ID does not exist" containerID="7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.002530 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099"} err="failed to get container status \"7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099\": rpc error: code = NotFound desc = could not find container \"7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099\": container with ID starting with 7e73331d80520e957a27267a2f94d941cc05d764871cf287425597849b62b099 not found: ID does not exist" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.002552 4706 scope.go:117] "RemoveContainer" containerID="0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98" Nov 27 07:13:19 crc kubenswrapper[4706]: E1127 07:13:19.002788 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98\": container with ID starting with 0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98 not found: ID does not exist" containerID="0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.002809 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98"} err="failed to get container status \"0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98\": rpc error: code = NotFound desc = could not find container \"0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98\": container with ID starting with 0794d8f314f702c19d06770b569ebb9c82d299f09ff684a0ddaed6f33c40ae98 not found: ID does not exist" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.002832 4706 scope.go:117] "RemoveContainer" containerID="00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75" Nov 27 07:13:19 crc kubenswrapper[4706]: E1127 07:13:19.003156 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75\": container with ID starting with 00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75 not found: ID does not exist" containerID="00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.003203 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75"} err="failed to get container status \"00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75\": rpc error: code = NotFound desc = could not find container \"00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75\": container with ID starting with 00cd338ab0e329ca1aeda60653a6da16fa5af7e9582b01feea1732ae1b684d75 not found: ID does not exist" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.003243 4706 scope.go:117] "RemoveContainer" containerID="0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.014357 4706 scope.go:117] "RemoveContainer" containerID="89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.027195 4706 scope.go:117] "RemoveContainer" containerID="8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.039552 4706 scope.go:117] "RemoveContainer" containerID="0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65" Nov 27 07:13:19 crc kubenswrapper[4706]: E1127 07:13:19.039966 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65\": container with ID starting with 0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65 not found: ID does not exist" containerID="0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.040005 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65"} err="failed to get container status \"0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65\": rpc error: code = NotFound desc = could not find container \"0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65\": container with ID starting with 0eb989cc7d57577a5f84ba33325d4942166d52d2756a5a4d1fcaa240aa9b7a65 not found: ID does not exist" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.040032 4706 scope.go:117] "RemoveContainer" containerID="89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785" Nov 27 07:13:19 crc kubenswrapper[4706]: E1127 07:13:19.040344 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785\": container with ID starting with 89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785 not found: ID does not exist" containerID="89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.040414 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785"} err="failed to get container status \"89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785\": rpc error: code = NotFound desc = could not find container \"89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785\": container with ID starting with 89237e8b1025516ec5d2b62eba3399ac0d5edc75b4fbe1dd1be85d55a83e0785 not found: ID does not exist" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.040442 4706 scope.go:117] "RemoveContainer" containerID="8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f" Nov 27 07:13:19 crc kubenswrapper[4706]: E1127 07:13:19.040757 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f\": container with ID starting with 8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f not found: ID does not exist" containerID="8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.040777 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f"} err="failed to get container status \"8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f\": rpc error: code = NotFound desc = could not find container \"8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f\": container with ID starting with 8217517047dc851e0a9ca954efa2b0770bb8b1a091893c27e651dc5a54d73d6f not found: ID does not exist" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.057474 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fzsrg"] Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.074812 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxm5z"] Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.080157 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxm5z"] Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.090970 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8vv7v"] Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.100410 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8vv7v"] Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.104061 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tq6jq"] Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.106557 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tq6jq"] Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.333278 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.334452 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.397910 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.397956 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.397997 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.398020 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.398031 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.398295 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.398918 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.398957 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.398984 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.407615 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.499411 4706 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.499460 4706 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.499472 4706 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.499490 4706 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.499502 4706 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.771504 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" event={"ID":"bd7353c5-075d-4f1d-b18e-2da68cc06652","Type":"ContainerStarted","Data":"86bcfaf656b37228fe46b0aefd72c5fc17a8d3c902761d61192d9e96cef93dee"} Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.771549 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" event={"ID":"bd7353c5-075d-4f1d-b18e-2da68cc06652","Type":"ContainerStarted","Data":"7c3da8d43fb77d6cd0e098b28644d3bbf45a2b8f3564dc4d204b3fd2a3707ae9"} Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.772679 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.777941 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.777993 4706 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa" exitCode=137 Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.778077 4706 scope.go:117] "RemoveContainer" containerID="ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.778189 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.779298 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.801402 4706 scope.go:117] "RemoveContainer" containerID="ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa" Nov 27 07:13:19 crc kubenswrapper[4706]: E1127 07:13:19.804532 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa\": container with ID starting with ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa not found: ID does not exist" containerID="ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.804572 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa"} err="failed to get container status \"ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa\": rpc error: code = NotFound desc = could not find container \"ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa\": container with ID starting with ba324e9c907641f3666456cfa09703a85a560913492502c9838ffa1cccb33baa not found: ID does not exist" Nov 27 07:13:19 crc kubenswrapper[4706]: I1127 07:13:19.805015 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fzsrg" podStartSLOduration=1.8049948580000001 podStartE2EDuration="1.804994858s" podCreationTimestamp="2025-11-27 07:13:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:13:19.792959679 +0000 UTC m=+283.682550499" watchObservedRunningTime="2025-11-27 07:13:19.804994858 +0000 UTC m=+283.694585688" Nov 27 07:13:20 crc kubenswrapper[4706]: I1127 07:13:20.786410 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" path="/var/lib/kubelet/pods/3b1176cb-10d1-4516-9ece-d51bb6ca515d/volumes" Nov 27 07:13:20 crc kubenswrapper[4706]: I1127 07:13:20.787557 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" path="/var/lib/kubelet/pods/43c8fce5-14bd-4d14-b488-50600b596cb5/volumes" Nov 27 07:13:20 crc kubenswrapper[4706]: I1127 07:13:20.788416 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" path="/var/lib/kubelet/pods/a8ac4a84-ddce-41e8-9da0-efd60117aa67/volumes" Nov 27 07:13:20 crc kubenswrapper[4706]: I1127 07:13:20.789468 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" path="/var/lib/kubelet/pods/adf72c72-f7cf-4ed2-9356-ef2c880d8164/volumes" Nov 27 07:13:20 crc kubenswrapper[4706]: I1127 07:13:20.790037 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" path="/var/lib/kubelet/pods/e22ae6d0-dbdf-4b69-84f0-0643266a7dff/volumes" Nov 27 07:13:20 crc kubenswrapper[4706]: I1127 07:13:20.790775 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.226173 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d2t2d"] Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.226989 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227003 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227014 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227020 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227029 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227037 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227045 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227052 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227061 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerName="marketplace-operator" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227067 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerName="marketplace-operator" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227077 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227083 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227092 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227097 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227106 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227112 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="extract-utilities" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227119 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227125 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227133 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227138 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227145 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227151 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227157 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227162 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: E1127 07:13:38.227170 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227175 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="extract-content" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227270 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c8fce5-14bd-4d14-b488-50600b596cb5" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227280 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8ac4a84-ddce-41e8-9da0-efd60117aa67" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227289 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1176cb-10d1-4516-9ece-d51bb6ca515d" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227296 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22ae6d0-dbdf-4b69-84f0-0643266a7dff" containerName="marketplace-operator" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227306 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf72c72-f7cf-4ed2-9356-ef2c880d8164" containerName="registry-server" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.227958 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.231501 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.235700 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2t2d"] Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.315271 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-catalog-content\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.315325 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-utilities\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.315348 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw9xz\" (UniqueName: \"kubernetes.io/projected/20f404ea-f279-4dc8-9a56-0cf44343ede6-kube-api-access-qw9xz\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.415981 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw9xz\" (UniqueName: \"kubernetes.io/projected/20f404ea-f279-4dc8-9a56-0cf44343ede6-kube-api-access-qw9xz\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.416065 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-catalog-content\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.416097 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-utilities\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.416534 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-utilities\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.416707 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-catalog-content\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.445889 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw9xz\" (UniqueName: \"kubernetes.io/projected/20f404ea-f279-4dc8-9a56-0cf44343ede6-kube-api-access-qw9xz\") pod \"community-operators-d2t2d\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.542519 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.737730 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2t2d"] Nov 27 07:13:38 crc kubenswrapper[4706]: W1127 07:13:38.744512 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20f404ea_f279_4dc8_9a56_0cf44343ede6.slice/crio-bbf35eedbf9f2a1d026d922aa2566ece573c8ccb850804accce4963be65bd981 WatchSource:0}: Error finding container bbf35eedbf9f2a1d026d922aa2566ece573c8ccb850804accce4963be65bd981: Status 404 returned error can't find the container with id bbf35eedbf9f2a1d026d922aa2566ece573c8ccb850804accce4963be65bd981 Nov 27 07:13:38 crc kubenswrapper[4706]: I1127 07:13:38.874577 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2t2d" event={"ID":"20f404ea-f279-4dc8-9a56-0cf44343ede6","Type":"ContainerStarted","Data":"bbf35eedbf9f2a1d026d922aa2566ece573c8ccb850804accce4963be65bd981"} Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.220781 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k7j8m"] Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.222296 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.224197 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.230533 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7j8m"] Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.326035 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c118696-a09a-405f-89e4-aedbe69487f1-utilities\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.326088 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tlvk\" (UniqueName: \"kubernetes.io/projected/6c118696-a09a-405f-89e4-aedbe69487f1-kube-api-access-6tlvk\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.326116 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c118696-a09a-405f-89e4-aedbe69487f1-catalog-content\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.427267 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c118696-a09a-405f-89e4-aedbe69487f1-utilities\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.427617 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tlvk\" (UniqueName: \"kubernetes.io/projected/6c118696-a09a-405f-89e4-aedbe69487f1-kube-api-access-6tlvk\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.427751 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c118696-a09a-405f-89e4-aedbe69487f1-catalog-content\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.427874 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c118696-a09a-405f-89e4-aedbe69487f1-utilities\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.428123 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c118696-a09a-405f-89e4-aedbe69487f1-catalog-content\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.449906 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tlvk\" (UniqueName: \"kubernetes.io/projected/6c118696-a09a-405f-89e4-aedbe69487f1-kube-api-access-6tlvk\") pod \"redhat-marketplace-k7j8m\" (UID: \"6c118696-a09a-405f-89e4-aedbe69487f1\") " pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.554288 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.791536 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7j8m"] Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.906885 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7j8m" event={"ID":"6c118696-a09a-405f-89e4-aedbe69487f1","Type":"ContainerStarted","Data":"b1de2ba1f767263ed841fdcc2ba264328e09fbf1bceedcc55cd98fb58c6a97bd"} Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.911298 4706 generic.go:334] "Generic (PLEG): container finished" podID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerID="57e4a22e7aa094044541569800f8c62a91c13c805e60f955a7db0fbc68a17660" exitCode=0 Nov 27 07:13:39 crc kubenswrapper[4706]: I1127 07:13:39.911327 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2t2d" event={"ID":"20f404ea-f279-4dc8-9a56-0cf44343ede6","Type":"ContainerDied","Data":"57e4a22e7aa094044541569800f8c62a91c13c805e60f955a7db0fbc68a17660"} Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.623392 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h4vk6"] Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.625888 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.627590 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.637865 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4vk6"] Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.643877 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547944fe-57e9-4370-90a3-9c60f4d3d03b-catalog-content\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.643931 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547944fe-57e9-4370-90a3-9c60f4d3d03b-utilities\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.643986 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw59l\" (UniqueName: \"kubernetes.io/projected/547944fe-57e9-4370-90a3-9c60f4d3d03b-kube-api-access-vw59l\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.745476 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547944fe-57e9-4370-90a3-9c60f4d3d03b-catalog-content\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.745816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547944fe-57e9-4370-90a3-9c60f4d3d03b-utilities\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.745864 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw59l\" (UniqueName: \"kubernetes.io/projected/547944fe-57e9-4370-90a3-9c60f4d3d03b-kube-api-access-vw59l\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.746356 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547944fe-57e9-4370-90a3-9c60f4d3d03b-catalog-content\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.746474 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547944fe-57e9-4370-90a3-9c60f4d3d03b-utilities\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.771841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw59l\" (UniqueName: \"kubernetes.io/projected/547944fe-57e9-4370-90a3-9c60f4d3d03b-kube-api-access-vw59l\") pod \"certified-operators-h4vk6\" (UID: \"547944fe-57e9-4370-90a3-9c60f4d3d03b\") " pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.916184 4706 generic.go:334] "Generic (PLEG): container finished" podID="6c118696-a09a-405f-89e4-aedbe69487f1" containerID="ae20df62f32d0383088a9432dc00059312c6e8b2dabd413f9d56c5db9f9dc600" exitCode=0 Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.916258 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7j8m" event={"ID":"6c118696-a09a-405f-89e4-aedbe69487f1","Type":"ContainerDied","Data":"ae20df62f32d0383088a9432dc00059312c6e8b2dabd413f9d56c5db9f9dc600"} Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.919161 4706 generic.go:334] "Generic (PLEG): container finished" podID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerID="f8acfae962b6171d0184004287668879c8c0e9d26a2d6a6d1a417335912c65bd" exitCode=0 Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.919191 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2t2d" event={"ID":"20f404ea-f279-4dc8-9a56-0cf44343ede6","Type":"ContainerDied","Data":"f8acfae962b6171d0184004287668879c8c0e9d26a2d6a6d1a417335912c65bd"} Nov 27 07:13:40 crc kubenswrapper[4706]: I1127 07:13:40.964372 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.155741 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4vk6"] Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.627451 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n274h"] Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.628594 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.631133 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.642046 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n274h"] Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.663716 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwhs2\" (UniqueName: \"kubernetes.io/projected/3b239571-088a-4b5f-ad24-39500fad1dea-kube-api-access-pwhs2\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.663763 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b239571-088a-4b5f-ad24-39500fad1dea-utilities\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.663796 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b239571-088a-4b5f-ad24-39500fad1dea-catalog-content\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.765302 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwhs2\" (UniqueName: \"kubernetes.io/projected/3b239571-088a-4b5f-ad24-39500fad1dea-kube-api-access-pwhs2\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.765365 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b239571-088a-4b5f-ad24-39500fad1dea-utilities\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.765411 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b239571-088a-4b5f-ad24-39500fad1dea-catalog-content\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.765895 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b239571-088a-4b5f-ad24-39500fad1dea-catalog-content\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.765954 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b239571-088a-4b5f-ad24-39500fad1dea-utilities\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.784003 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwhs2\" (UniqueName: \"kubernetes.io/projected/3b239571-088a-4b5f-ad24-39500fad1dea-kube-api-access-pwhs2\") pod \"redhat-operators-n274h\" (UID: \"3b239571-088a-4b5f-ad24-39500fad1dea\") " pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.925186 4706 generic.go:334] "Generic (PLEG): container finished" podID="6c118696-a09a-405f-89e4-aedbe69487f1" containerID="9fc766c732b466d10d449c60c987f506f53186c0e8851325d74559a4a1ed85e0" exitCode=0 Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.925256 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7j8m" event={"ID":"6c118696-a09a-405f-89e4-aedbe69487f1","Type":"ContainerDied","Data":"9fc766c732b466d10d449c60c987f506f53186c0e8851325d74559a4a1ed85e0"} Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.926497 4706 generic.go:334] "Generic (PLEG): container finished" podID="547944fe-57e9-4370-90a3-9c60f4d3d03b" containerID="9e213d247c8adbe0a8c88f2d6d3682f62b1259025e2a3d49402ba33c35c61518" exitCode=0 Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.926572 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vk6" event={"ID":"547944fe-57e9-4370-90a3-9c60f4d3d03b","Type":"ContainerDied","Data":"9e213d247c8adbe0a8c88f2d6d3682f62b1259025e2a3d49402ba33c35c61518"} Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.926615 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vk6" event={"ID":"547944fe-57e9-4370-90a3-9c60f4d3d03b","Type":"ContainerStarted","Data":"b68b693ce597f6d11ddeb58864df812f10482452d7e2fe847c0e88b03b356a01"} Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.928631 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2t2d" event={"ID":"20f404ea-f279-4dc8-9a56-0cf44343ede6","Type":"ContainerStarted","Data":"6c694b87d3e240f20100347fc67d101de4741cab8a4788abc2d431559a212d51"} Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.976106 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d2t2d" podStartSLOduration=2.474616923 podStartE2EDuration="3.976087608s" podCreationTimestamp="2025-11-27 07:13:38 +0000 UTC" firstStartedPulling="2025-11-27 07:13:39.912764715 +0000 UTC m=+303.802355515" lastFinishedPulling="2025-11-27 07:13:41.41423539 +0000 UTC m=+305.303826200" observedRunningTime="2025-11-27 07:13:41.975906702 +0000 UTC m=+305.865497512" watchObservedRunningTime="2025-11-27 07:13:41.976087608 +0000 UTC m=+305.865678418" Nov 27 07:13:41 crc kubenswrapper[4706]: I1127 07:13:41.976252 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.164782 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n274h"] Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.169189 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lbldb"] Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.169462 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" podUID="50c0d8f5-2115-4b0c-957d-9e687cdb1d57" containerName="controller-manager" containerID="cri-o://382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2" gracePeriod=30 Nov 27 07:13:42 crc kubenswrapper[4706]: W1127 07:13:42.174096 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b239571_088a_4b5f_ad24_39500fad1dea.slice/crio-831d7b64f95b2781e23410986bde1e784e80cf6498cc8852c513ffbf78440806 WatchSource:0}: Error finding container 831d7b64f95b2781e23410986bde1e784e80cf6498cc8852c513ffbf78440806: Status 404 returned error can't find the container with id 831d7b64f95b2781e23410986bde1e784e80cf6498cc8852c513ffbf78440806 Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.264752 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k"] Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.264985 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" podUID="a0558cfe-6d59-4897-9a10-595f42734ca6" containerName="route-controller-manager" containerID="cri-o://0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f" gracePeriod=30 Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.476004 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.573085 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-client-ca\") pod \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.573126 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwj8d\" (UniqueName: \"kubernetes.io/projected/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-kube-api-access-xwj8d\") pod \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.573148 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-serving-cert\") pod \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.573189 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-proxy-ca-bundles\") pod \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.573210 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-config\") pod \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\" (UID: \"50c0d8f5-2115-4b0c-957d-9e687cdb1d57\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.574165 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-config" (OuterVolumeSpecName: "config") pod "50c0d8f5-2115-4b0c-957d-9e687cdb1d57" (UID: "50c0d8f5-2115-4b0c-957d-9e687cdb1d57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.574568 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-client-ca" (OuterVolumeSpecName: "client-ca") pod "50c0d8f5-2115-4b0c-957d-9e687cdb1d57" (UID: "50c0d8f5-2115-4b0c-957d-9e687cdb1d57"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.574825 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "50c0d8f5-2115-4b0c-957d-9e687cdb1d57" (UID: "50c0d8f5-2115-4b0c-957d-9e687cdb1d57"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.578372 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "50c0d8f5-2115-4b0c-957d-9e687cdb1d57" (UID: "50c0d8f5-2115-4b0c-957d-9e687cdb1d57"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.578434 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-kube-api-access-xwj8d" (OuterVolumeSpecName: "kube-api-access-xwj8d") pod "50c0d8f5-2115-4b0c-957d-9e687cdb1d57" (UID: "50c0d8f5-2115-4b0c-957d-9e687cdb1d57"). InnerVolumeSpecName "kube-api-access-xwj8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.616434 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.673984 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.674015 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.674025 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.674037 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwj8d\" (UniqueName: \"kubernetes.io/projected/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-kube-api-access-xwj8d\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.674046 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50c0d8f5-2115-4b0c-957d-9e687cdb1d57-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.774431 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-config\") pod \"a0558cfe-6d59-4897-9a10-595f42734ca6\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.774492 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k299h\" (UniqueName: \"kubernetes.io/projected/a0558cfe-6d59-4897-9a10-595f42734ca6-kube-api-access-k299h\") pod \"a0558cfe-6d59-4897-9a10-595f42734ca6\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.774529 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0558cfe-6d59-4897-9a10-595f42734ca6-serving-cert\") pod \"a0558cfe-6d59-4897-9a10-595f42734ca6\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.774549 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-client-ca\") pod \"a0558cfe-6d59-4897-9a10-595f42734ca6\" (UID: \"a0558cfe-6d59-4897-9a10-595f42734ca6\") " Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.775360 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-client-ca" (OuterVolumeSpecName: "client-ca") pod "a0558cfe-6d59-4897-9a10-595f42734ca6" (UID: "a0558cfe-6d59-4897-9a10-595f42734ca6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.775415 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-config" (OuterVolumeSpecName: "config") pod "a0558cfe-6d59-4897-9a10-595f42734ca6" (UID: "a0558cfe-6d59-4897-9a10-595f42734ca6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.778852 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0558cfe-6d59-4897-9a10-595f42734ca6-kube-api-access-k299h" (OuterVolumeSpecName: "kube-api-access-k299h") pod "a0558cfe-6d59-4897-9a10-595f42734ca6" (UID: "a0558cfe-6d59-4897-9a10-595f42734ca6"). InnerVolumeSpecName "kube-api-access-k299h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.793674 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0558cfe-6d59-4897-9a10-595f42734ca6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a0558cfe-6d59-4897-9a10-595f42734ca6" (UID: "a0558cfe-6d59-4897-9a10-595f42734ca6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.876984 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.877025 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k299h\" (UniqueName: \"kubernetes.io/projected/a0558cfe-6d59-4897-9a10-595f42734ca6-kube-api-access-k299h\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.877037 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0558cfe-6d59-4897-9a10-595f42734ca6-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.877046 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0558cfe-6d59-4897-9a10-595f42734ca6-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.935148 4706 generic.go:334] "Generic (PLEG): container finished" podID="50c0d8f5-2115-4b0c-957d-9e687cdb1d57" containerID="382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2" exitCode=0 Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.935334 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" event={"ID":"50c0d8f5-2115-4b0c-957d-9e687cdb1d57","Type":"ContainerDied","Data":"382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.935560 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" event={"ID":"50c0d8f5-2115-4b0c-957d-9e687cdb1d57","Type":"ContainerDied","Data":"fa6704ca2eebb36a1ec03614f4f92d42f69a61126a66bc3cec80d4d3e52342ea"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.935577 4706 scope.go:117] "RemoveContainer" containerID="382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.935381 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lbldb" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.937666 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b239571-088a-4b5f-ad24-39500fad1dea" containerID="51952f36fdddaeba1d52faaa423a083f05913fe8b3abc2eb6397a714f4e332f8" exitCode=0 Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.937719 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n274h" event={"ID":"3b239571-088a-4b5f-ad24-39500fad1dea","Type":"ContainerDied","Data":"51952f36fdddaeba1d52faaa423a083f05913fe8b3abc2eb6397a714f4e332f8"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.937740 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n274h" event={"ID":"3b239571-088a-4b5f-ad24-39500fad1dea","Type":"ContainerStarted","Data":"831d7b64f95b2781e23410986bde1e784e80cf6498cc8852c513ffbf78440806"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.942196 4706 generic.go:334] "Generic (PLEG): container finished" podID="a0558cfe-6d59-4897-9a10-595f42734ca6" containerID="0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f" exitCode=0 Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.942322 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.943108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" event={"ID":"a0558cfe-6d59-4897-9a10-595f42734ca6","Type":"ContainerDied","Data":"0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.943136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k" event={"ID":"a0558cfe-6d59-4897-9a10-595f42734ca6","Type":"ContainerDied","Data":"4a5f70accc9925d8d4df8196684f11e0ad9e5f771785a755f7e5d8373cd24c96"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.945503 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7j8m" event={"ID":"6c118696-a09a-405f-89e4-aedbe69487f1","Type":"ContainerStarted","Data":"e705948db202545a49f9d1c416eda5c51c3c589d4ac3270de8de352cefecf444"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.953743 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vk6" event={"ID":"547944fe-57e9-4370-90a3-9c60f4d3d03b","Type":"ContainerStarted","Data":"ef396dd8b00b1de78f8625736e375d413c58af160912933ed0ddfc3020fe8ca8"} Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.957112 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lbldb"] Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.958694 4706 scope.go:117] "RemoveContainer" containerID="382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2" Nov 27 07:13:42 crc kubenswrapper[4706]: E1127 07:13:42.962326 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2\": container with ID starting with 382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2 not found: ID does not exist" containerID="382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.962361 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2"} err="failed to get container status \"382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2\": rpc error: code = NotFound desc = could not find container \"382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2\": container with ID starting with 382c7d1ff7caaf34ef13569c7f51381c8d3006777ea2108bd65e8e01099995a2 not found: ID does not exist" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.962380 4706 scope.go:117] "RemoveContainer" containerID="0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.962617 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lbldb"] Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.980060 4706 scope.go:117] "RemoveContainer" containerID="0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f" Nov 27 07:13:42 crc kubenswrapper[4706]: E1127 07:13:42.980413 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f\": container with ID starting with 0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f not found: ID does not exist" containerID="0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.980444 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f"} err="failed to get container status \"0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f\": rpc error: code = NotFound desc = could not find container \"0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f\": container with ID starting with 0f14d1e1486156c8315a6825ea3b48d8728386290a990c44df889d39ec130f3f not found: ID does not exist" Nov 27 07:13:42 crc kubenswrapper[4706]: I1127 07:13:42.994758 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k7j8m" podStartSLOduration=2.46291561 podStartE2EDuration="3.994740789s" podCreationTimestamp="2025-11-27 07:13:39 +0000 UTC" firstStartedPulling="2025-11-27 07:13:40.91733965 +0000 UTC m=+304.806930460" lastFinishedPulling="2025-11-27 07:13:42.449164829 +0000 UTC m=+306.338755639" observedRunningTime="2025-11-27 07:13:42.990064464 +0000 UTC m=+306.879655274" watchObservedRunningTime="2025-11-27 07:13:42.994740789 +0000 UTC m=+306.884331599" Nov 27 07:13:43 crc kubenswrapper[4706]: I1127 07:13:43.015227 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k"] Nov 27 07:13:43 crc kubenswrapper[4706]: I1127 07:13:43.018206 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q7d5k"] Nov 27 07:13:43 crc kubenswrapper[4706]: I1127 07:13:43.962075 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n274h" event={"ID":"3b239571-088a-4b5f-ad24-39500fad1dea","Type":"ContainerStarted","Data":"8048cbc615813ba9147dc7e874662a1589e4bfaa7b790c2161aff28588165fb7"} Nov 27 07:13:43 crc kubenswrapper[4706]: I1127 07:13:43.966804 4706 generic.go:334] "Generic (PLEG): container finished" podID="547944fe-57e9-4370-90a3-9c60f4d3d03b" containerID="ef396dd8b00b1de78f8625736e375d413c58af160912933ed0ddfc3020fe8ca8" exitCode=0 Nov 27 07:13:43 crc kubenswrapper[4706]: I1127 07:13:43.966895 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vk6" event={"ID":"547944fe-57e9-4370-90a3-9c60f4d3d03b","Type":"ContainerDied","Data":"ef396dd8b00b1de78f8625736e375d413c58af160912933ed0ddfc3020fe8ca8"} Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.039633 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq"] Nov 27 07:13:44 crc kubenswrapper[4706]: E1127 07:13:44.039889 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c0d8f5-2115-4b0c-957d-9e687cdb1d57" containerName="controller-manager" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.039904 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c0d8f5-2115-4b0c-957d-9e687cdb1d57" containerName="controller-manager" Nov 27 07:13:44 crc kubenswrapper[4706]: E1127 07:13:44.039916 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0558cfe-6d59-4897-9a10-595f42734ca6" containerName="route-controller-manager" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.039923 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0558cfe-6d59-4897-9a10-595f42734ca6" containerName="route-controller-manager" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.040041 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c0d8f5-2115-4b0c-957d-9e687cdb1d57" containerName="controller-manager" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.040060 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0558cfe-6d59-4897-9a10-595f42734ca6" containerName="route-controller-manager" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.040578 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.042820 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-64f6dbf577-fr495"] Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.044364 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.044913 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.046003 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.046141 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.046305 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.046853 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.047052 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.053507 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.054259 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.054314 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.054555 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.054698 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.055793 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq"] Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.057658 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.058414 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.062391 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64f6dbf577-fr495"] Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091289 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-client-ca\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091330 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-client-ca\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091346 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-config\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091367 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-config\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091382 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-serving-cert\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091454 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctt95\" (UniqueName: \"kubernetes.io/projected/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-kube-api-access-ctt95\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091471 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4aca52a-365b-4ce3-9328-fec5a286eb4e-serving-cert\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091491 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cwsr\" (UniqueName: \"kubernetes.io/projected/c4aca52a-365b-4ce3-9328-fec5a286eb4e-kube-api-access-4cwsr\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.091537 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-proxy-ca-bundles\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192085 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctt95\" (UniqueName: \"kubernetes.io/projected/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-kube-api-access-ctt95\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192140 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4aca52a-365b-4ce3-9328-fec5a286eb4e-serving-cert\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192171 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cwsr\" (UniqueName: \"kubernetes.io/projected/c4aca52a-365b-4ce3-9328-fec5a286eb4e-kube-api-access-4cwsr\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192197 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-proxy-ca-bundles\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192247 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-client-ca\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192271 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-client-ca\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192292 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-config\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192319 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-config\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.192343 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-serving-cert\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.193216 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-client-ca\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.193281 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-client-ca\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.193700 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-config\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.193788 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-config\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.194435 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-proxy-ca-bundles\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.200069 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4aca52a-365b-4ce3-9328-fec5a286eb4e-serving-cert\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.200081 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-serving-cert\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.212905 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctt95\" (UniqueName: \"kubernetes.io/projected/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-kube-api-access-ctt95\") pod \"controller-manager-64f6dbf577-fr495\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.212956 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cwsr\" (UniqueName: \"kubernetes.io/projected/c4aca52a-365b-4ce3-9328-fec5a286eb4e-kube-api-access-4cwsr\") pod \"route-controller-manager-797fd966f5-hb6sq\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.368479 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.379348 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.535806 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq"] Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.579884 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64f6dbf577-fr495"] Nov 27 07:13:44 crc kubenswrapper[4706]: W1127 07:13:44.583209 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0bdca0d_931f_4a7e_a9cf_464798ab9ab7.slice/crio-4938ae5b2a28b813c4f293fe295a80aa0342c863dfd8d9e634429f7ec4d54981 WatchSource:0}: Error finding container 4938ae5b2a28b813c4f293fe295a80aa0342c863dfd8d9e634429f7ec4d54981: Status 404 returned error can't find the container with id 4938ae5b2a28b813c4f293fe295a80aa0342c863dfd8d9e634429f7ec4d54981 Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.783645 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c0d8f5-2115-4b0c-957d-9e687cdb1d57" path="/var/lib/kubelet/pods/50c0d8f5-2115-4b0c-957d-9e687cdb1d57/volumes" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.784312 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0558cfe-6d59-4897-9a10-595f42734ca6" path="/var/lib/kubelet/pods/a0558cfe-6d59-4897-9a10-595f42734ca6/volumes" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.973406 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" event={"ID":"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7","Type":"ContainerStarted","Data":"442aaf751bf15656804c06d234c6c92a7aae2dc5ed7ac420dd3d9861eef68e6f"} Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.973447 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" event={"ID":"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7","Type":"ContainerStarted","Data":"4938ae5b2a28b813c4f293fe295a80aa0342c863dfd8d9e634429f7ec4d54981"} Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.973621 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.977001 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vk6" event={"ID":"547944fe-57e9-4370-90a3-9c60f4d3d03b","Type":"ContainerStarted","Data":"32c976bc579c4effeee183807febc68f13bd05d283e116c5188a3c6a8f51ad94"} Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.979010 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b239571-088a-4b5f-ad24-39500fad1dea" containerID="8048cbc615813ba9147dc7e874662a1589e4bfaa7b790c2161aff28588165fb7" exitCode=0 Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.979072 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n274h" event={"ID":"3b239571-088a-4b5f-ad24-39500fad1dea","Type":"ContainerDied","Data":"8048cbc615813ba9147dc7e874662a1589e4bfaa7b790c2161aff28588165fb7"} Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.981029 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" event={"ID":"c4aca52a-365b-4ce3-9328-fec5a286eb4e","Type":"ContainerStarted","Data":"2a7f7040cdd18ce4d967dad34643b3482828400a49b1427dba1ba48d71e996eb"} Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.981068 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" event={"ID":"c4aca52a-365b-4ce3-9328-fec5a286eb4e","Type":"ContainerStarted","Data":"0eb4969e6d1ea448d081dccee6bcd6f514ff6ad87c2678c94b2feef1e087e208"} Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.981253 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:44 crc kubenswrapper[4706]: I1127 07:13:44.986460 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:45 crc kubenswrapper[4706]: I1127 07:13:45.008363 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" podStartSLOduration=3.008340377 podStartE2EDuration="3.008340377s" podCreationTimestamp="2025-11-27 07:13:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:13:45.00600524 +0000 UTC m=+308.895596050" watchObservedRunningTime="2025-11-27 07:13:45.008340377 +0000 UTC m=+308.897931187" Nov 27 07:13:45 crc kubenswrapper[4706]: I1127 07:13:45.024021 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h4vk6" podStartSLOduration=2.411866115 podStartE2EDuration="5.024002405s" podCreationTimestamp="2025-11-27 07:13:40 +0000 UTC" firstStartedPulling="2025-11-27 07:13:41.928786753 +0000 UTC m=+305.818377563" lastFinishedPulling="2025-11-27 07:13:44.540923043 +0000 UTC m=+308.430513853" observedRunningTime="2025-11-27 07:13:45.022972231 +0000 UTC m=+308.912563061" watchObservedRunningTime="2025-11-27 07:13:45.024002405 +0000 UTC m=+308.913593215" Nov 27 07:13:45 crc kubenswrapper[4706]: I1127 07:13:45.096414 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" podStartSLOduration=3.096395621 podStartE2EDuration="3.096395621s" podCreationTimestamp="2025-11-27 07:13:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:13:45.093952799 +0000 UTC m=+308.983543609" watchObservedRunningTime="2025-11-27 07:13:45.096395621 +0000 UTC m=+308.985986431" Nov 27 07:13:45 crc kubenswrapper[4706]: I1127 07:13:45.105912 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:45 crc kubenswrapper[4706]: I1127 07:13:45.989951 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n274h" event={"ID":"3b239571-088a-4b5f-ad24-39500fad1dea","Type":"ContainerStarted","Data":"87a0a7a0d26ea8cd1f752b6768e756bbe1acfcb2036112b38ee9f7997e98141c"} Nov 27 07:13:46 crc kubenswrapper[4706]: I1127 07:13:46.007368 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n274h" podStartSLOduration=2.4567440830000002 podStartE2EDuration="5.007352838s" podCreationTimestamp="2025-11-27 07:13:41 +0000 UTC" firstStartedPulling="2025-11-27 07:13:42.93916821 +0000 UTC m=+306.828759020" lastFinishedPulling="2025-11-27 07:13:45.489776965 +0000 UTC m=+309.379367775" observedRunningTime="2025-11-27 07:13:46.005120495 +0000 UTC m=+309.894711305" watchObservedRunningTime="2025-11-27 07:13:46.007352838 +0000 UTC m=+309.896943648" Nov 27 07:13:48 crc kubenswrapper[4706]: I1127 07:13:48.542954 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:48 crc kubenswrapper[4706]: I1127 07:13:48.543433 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:48 crc kubenswrapper[4706]: I1127 07:13:48.590863 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64f6dbf577-fr495"] Nov 27 07:13:48 crc kubenswrapper[4706]: I1127 07:13:48.591313 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" podUID="c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" containerName="controller-manager" containerID="cri-o://442aaf751bf15656804c06d234c6c92a7aae2dc5ed7ac420dd3d9861eef68e6f" gracePeriod=30 Nov 27 07:13:48 crc kubenswrapper[4706]: I1127 07:13:48.597467 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:48 crc kubenswrapper[4706]: I1127 07:13:48.613672 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq"] Nov 27 07:13:48 crc kubenswrapper[4706]: I1127 07:13:48.613894 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" podUID="c4aca52a-365b-4ce3-9328-fec5a286eb4e" containerName="route-controller-manager" containerID="cri-o://2a7f7040cdd18ce4d967dad34643b3482828400a49b1427dba1ba48d71e996eb" gracePeriod=30 Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.003789 4706 generic.go:334] "Generic (PLEG): container finished" podID="c4aca52a-365b-4ce3-9328-fec5a286eb4e" containerID="2a7f7040cdd18ce4d967dad34643b3482828400a49b1427dba1ba48d71e996eb" exitCode=0 Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.003879 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" event={"ID":"c4aca52a-365b-4ce3-9328-fec5a286eb4e","Type":"ContainerDied","Data":"2a7f7040cdd18ce4d967dad34643b3482828400a49b1427dba1ba48d71e996eb"} Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.005128 4706 generic.go:334] "Generic (PLEG): container finished" podID="c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" containerID="442aaf751bf15656804c06d234c6c92a7aae2dc5ed7ac420dd3d9861eef68e6f" exitCode=0 Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.005200 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" event={"ID":"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7","Type":"ContainerDied","Data":"442aaf751bf15656804c06d234c6c92a7aae2dc5ed7ac420dd3d9861eef68e6f"} Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.055788 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.102067 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.105212 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249206 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-proxy-ca-bundles\") pod \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249270 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-client-ca\") pod \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249302 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-client-ca\") pod \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249359 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-config\") pod \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-config\") pod \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249401 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4aca52a-365b-4ce3-9328-fec5a286eb4e-serving-cert\") pod \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249429 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cwsr\" (UniqueName: \"kubernetes.io/projected/c4aca52a-365b-4ce3-9328-fec5a286eb4e-kube-api-access-4cwsr\") pod \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\" (UID: \"c4aca52a-365b-4ce3-9328-fec5a286eb4e\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249451 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctt95\" (UniqueName: \"kubernetes.io/projected/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-kube-api-access-ctt95\") pod \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.249485 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-serving-cert\") pod \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\" (UID: \"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7\") " Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.251011 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-config" (OuterVolumeSpecName: "config") pod "c4aca52a-365b-4ce3-9328-fec5a286eb4e" (UID: "c4aca52a-365b-4ce3-9328-fec5a286eb4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.251047 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-client-ca" (OuterVolumeSpecName: "client-ca") pod "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" (UID: "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.251113 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" (UID: "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.251600 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-client-ca" (OuterVolumeSpecName: "client-ca") pod "c4aca52a-365b-4ce3-9328-fec5a286eb4e" (UID: "c4aca52a-365b-4ce3-9328-fec5a286eb4e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.251854 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-config" (OuterVolumeSpecName: "config") pod "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" (UID: "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.255004 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4aca52a-365b-4ce3-9328-fec5a286eb4e-kube-api-access-4cwsr" (OuterVolumeSpecName: "kube-api-access-4cwsr") pod "c4aca52a-365b-4ce3-9328-fec5a286eb4e" (UID: "c4aca52a-365b-4ce3-9328-fec5a286eb4e"). InnerVolumeSpecName "kube-api-access-4cwsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.255039 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4aca52a-365b-4ce3-9328-fec5a286eb4e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c4aca52a-365b-4ce3-9328-fec5a286eb4e" (UID: "c4aca52a-365b-4ce3-9328-fec5a286eb4e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.255145 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" (UID: "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.256429 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-kube-api-access-ctt95" (OuterVolumeSpecName: "kube-api-access-ctt95") pod "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" (UID: "c0bdca0d-931f-4a7e-a9cf-464798ab9ab7"). InnerVolumeSpecName "kube-api-access-ctt95". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350791 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350828 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350838 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4aca52a-365b-4ce3-9328-fec5a286eb4e-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350848 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cwsr\" (UniqueName: \"kubernetes.io/projected/c4aca52a-365b-4ce3-9328-fec5a286eb4e-kube-api-access-4cwsr\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350860 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctt95\" (UniqueName: \"kubernetes.io/projected/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-kube-api-access-ctt95\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350867 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350875 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350882 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.350890 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4aca52a-365b-4ce3-9328-fec5a286eb4e-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.555032 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.555080 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.592781 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.647709 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6c7745c94d-n92jt"] Nov 27 07:13:49 crc kubenswrapper[4706]: E1127 07:13:49.648204 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" containerName="controller-manager" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.648321 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" containerName="controller-manager" Nov 27 07:13:49 crc kubenswrapper[4706]: E1127 07:13:49.648397 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4aca52a-365b-4ce3-9328-fec5a286eb4e" containerName="route-controller-manager" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.648458 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4aca52a-365b-4ce3-9328-fec5a286eb4e" containerName="route-controller-manager" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.648594 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" containerName="controller-manager" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.648658 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4aca52a-365b-4ce3-9328-fec5a286eb4e" containerName="route-controller-manager" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.649069 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.656714 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c7745c94d-n92jt"] Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.755494 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-config\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.755806 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee31dce3-0174-49a2-afce-4cb0e5838256-serving-cert\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.755923 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-client-ca\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.756022 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-proxy-ca-bundles\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.756115 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qdpt\" (UniqueName: \"kubernetes.io/projected/ee31dce3-0174-49a2-afce-4cb0e5838256-kube-api-access-6qdpt\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.763246 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6"] Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.764003 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.773726 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6"] Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.857831 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-config\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.857883 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee31dce3-0174-49a2-afce-4cb0e5838256-serving-cert\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.857924 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-client-ca\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.857962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cef8686-b75a-4164-ab52-c3500ed736ef-serving-cert\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.857990 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jklwg\" (UniqueName: \"kubernetes.io/projected/8cef8686-b75a-4164-ab52-c3500ed736ef-kube-api-access-jklwg\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.858012 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-proxy-ca-bundles\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.858049 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qdpt\" (UniqueName: \"kubernetes.io/projected/ee31dce3-0174-49a2-afce-4cb0e5838256-kube-api-access-6qdpt\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.858073 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-client-ca\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.858117 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-config\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.859131 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-client-ca\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.859269 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-proxy-ca-bundles\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.859747 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-config\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.862078 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee31dce3-0174-49a2-afce-4cb0e5838256-serving-cert\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.890702 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qdpt\" (UniqueName: \"kubernetes.io/projected/ee31dce3-0174-49a2-afce-4cb0e5838256-kube-api-access-6qdpt\") pod \"controller-manager-6c7745c94d-n92jt\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.958696 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-client-ca\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.958753 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-config\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.958806 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cef8686-b75a-4164-ab52-c3500ed736ef-serving-cert\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.958824 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jklwg\" (UniqueName: \"kubernetes.io/projected/8cef8686-b75a-4164-ab52-c3500ed736ef-kube-api-access-jklwg\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.959982 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-client-ca\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.960375 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-config\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.962856 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cef8686-b75a-4164-ab52-c3500ed736ef-serving-cert\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:49 crc kubenswrapper[4706]: I1127 07:13:49.977690 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jklwg\" (UniqueName: \"kubernetes.io/projected/8cef8686-b75a-4164-ab52-c3500ed736ef-kube-api-access-jklwg\") pod \"route-controller-manager-696c5bc4d6-lsdl6\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.010817 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" event={"ID":"c0bdca0d-931f-4a7e-a9cf-464798ab9ab7","Type":"ContainerDied","Data":"4938ae5b2a28b813c4f293fe295a80aa0342c863dfd8d9e634429f7ec4d54981"} Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.010843 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64f6dbf577-fr495" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.010865 4706 scope.go:117] "RemoveContainer" containerID="442aaf751bf15656804c06d234c6c92a7aae2dc5ed7ac420dd3d9861eef68e6f" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.011461 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.014340 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" event={"ID":"c4aca52a-365b-4ce3-9328-fec5a286eb4e","Type":"ContainerDied","Data":"0eb4969e6d1ea448d081dccee6bcd6f514ff6ad87c2678c94b2feef1e087e208"} Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.014404 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.042352 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64f6dbf577-fr495"] Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.044752 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-64f6dbf577-fr495"] Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.045193 4706 scope.go:117] "RemoveContainer" containerID="2a7f7040cdd18ce4d967dad34643b3482828400a49b1427dba1ba48d71e996eb" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.056944 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq"] Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.060925 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797fd966f5-hb6sq"] Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.065345 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k7j8m" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.090263 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.249324 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c7745c94d-n92jt"] Nov 27 07:13:50 crc kubenswrapper[4706]: W1127 07:13:50.255472 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee31dce3_0174_49a2_afce_4cb0e5838256.slice/crio-34daa25d2218ee26a3e47bd23bee95a76e6a6257e0243e9a6db6b53aa296d7f1 WatchSource:0}: Error finding container 34daa25d2218ee26a3e47bd23bee95a76e6a6257e0243e9a6db6b53aa296d7f1: Status 404 returned error can't find the container with id 34daa25d2218ee26a3e47bd23bee95a76e6a6257e0243e9a6db6b53aa296d7f1 Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.339297 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6"] Nov 27 07:13:50 crc kubenswrapper[4706]: W1127 07:13:50.356331 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cef8686_b75a_4164_ab52_c3500ed736ef.slice/crio-0181d14a94d60067d348a54dcf78b61eaf0be42b5054283ea2bd73328d025d84 WatchSource:0}: Error finding container 0181d14a94d60067d348a54dcf78b61eaf0be42b5054283ea2bd73328d025d84: Status 404 returned error can't find the container with id 0181d14a94d60067d348a54dcf78b61eaf0be42b5054283ea2bd73328d025d84 Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.788751 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0bdca0d-931f-4a7e-a9cf-464798ab9ab7" path="/var/lib/kubelet/pods/c0bdca0d-931f-4a7e-a9cf-464798ab9ab7/volumes" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.789589 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4aca52a-365b-4ce3-9328-fec5a286eb4e" path="/var/lib/kubelet/pods/c4aca52a-365b-4ce3-9328-fec5a286eb4e/volumes" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.965067 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:50 crc kubenswrapper[4706]: I1127 07:13:50.965127 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.003501 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.020320 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" event={"ID":"8cef8686-b75a-4164-ab52-c3500ed736ef","Type":"ContainerStarted","Data":"49ed01abacbfe1fba1a47421ae6456e9062242a96fb7957db1c08aa802d20f00"} Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.020371 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" event={"ID":"8cef8686-b75a-4164-ab52-c3500ed736ef","Type":"ContainerStarted","Data":"0181d14a94d60067d348a54dcf78b61eaf0be42b5054283ea2bd73328d025d84"} Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.020711 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.025801 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" event={"ID":"ee31dce3-0174-49a2-afce-4cb0e5838256","Type":"ContainerStarted","Data":"5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4"} Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.025846 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.025856 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" event={"ID":"ee31dce3-0174-49a2-afce-4cb0e5838256","Type":"ContainerStarted","Data":"34daa25d2218ee26a3e47bd23bee95a76e6a6257e0243e9a6db6b53aa296d7f1"} Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.029946 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.030333 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.040513 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" podStartSLOduration=2.040500845 podStartE2EDuration="2.040500845s" podCreationTimestamp="2025-11-27 07:13:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:13:51.04001798 +0000 UTC m=+314.929608790" watchObservedRunningTime="2025-11-27 07:13:51.040500845 +0000 UTC m=+314.930091645" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.070302 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" podStartSLOduration=2.070282541 podStartE2EDuration="2.070282541s" podCreationTimestamp="2025-11-27 07:13:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:13:51.068052927 +0000 UTC m=+314.957643737" watchObservedRunningTime="2025-11-27 07:13:51.070282541 +0000 UTC m=+314.959873361" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.075758 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h4vk6" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.976752 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:51 crc kubenswrapper[4706]: I1127 07:13:51.977084 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:52 crc kubenswrapper[4706]: I1127 07:13:52.018021 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:13:52 crc kubenswrapper[4706]: I1127 07:13:52.065987 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n274h" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.141507 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c7745c94d-n92jt"] Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.142517 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" podUID="ee31dce3-0174-49a2-afce-4cb0e5838256" containerName="controller-manager" containerID="cri-o://5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4" gracePeriod=30 Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.652085 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.810190 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-client-ca\") pod \"ee31dce3-0174-49a2-afce-4cb0e5838256\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.810266 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-config\") pod \"ee31dce3-0174-49a2-afce-4cb0e5838256\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.810380 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qdpt\" (UniqueName: \"kubernetes.io/projected/ee31dce3-0174-49a2-afce-4cb0e5838256-kube-api-access-6qdpt\") pod \"ee31dce3-0174-49a2-afce-4cb0e5838256\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.810407 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-proxy-ca-bundles\") pod \"ee31dce3-0174-49a2-afce-4cb0e5838256\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.810434 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee31dce3-0174-49a2-afce-4cb0e5838256-serving-cert\") pod \"ee31dce3-0174-49a2-afce-4cb0e5838256\" (UID: \"ee31dce3-0174-49a2-afce-4cb0e5838256\") " Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.811119 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-client-ca" (OuterVolumeSpecName: "client-ca") pod "ee31dce3-0174-49a2-afce-4cb0e5838256" (UID: "ee31dce3-0174-49a2-afce-4cb0e5838256"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.811272 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ee31dce3-0174-49a2-afce-4cb0e5838256" (UID: "ee31dce3-0174-49a2-afce-4cb0e5838256"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.811779 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-config" (OuterVolumeSpecName: "config") pod "ee31dce3-0174-49a2-afce-4cb0e5838256" (UID: "ee31dce3-0174-49a2-afce-4cb0e5838256"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.815890 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee31dce3-0174-49a2-afce-4cb0e5838256-kube-api-access-6qdpt" (OuterVolumeSpecName: "kube-api-access-6qdpt") pod "ee31dce3-0174-49a2-afce-4cb0e5838256" (UID: "ee31dce3-0174-49a2-afce-4cb0e5838256"). InnerVolumeSpecName "kube-api-access-6qdpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.819409 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee31dce3-0174-49a2-afce-4cb0e5838256-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ee31dce3-0174-49a2-afce-4cb0e5838256" (UID: "ee31dce3-0174-49a2-afce-4cb0e5838256"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.911625 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qdpt\" (UniqueName: \"kubernetes.io/projected/ee31dce3-0174-49a2-afce-4cb0e5838256-kube-api-access-6qdpt\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.911661 4706 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.911671 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee31dce3-0174-49a2-afce-4cb0e5838256-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.911682 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:02 crc kubenswrapper[4706]: I1127 07:14:02.911691 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee31dce3-0174-49a2-afce-4cb0e5838256-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.078838 4706 generic.go:334] "Generic (PLEG): container finished" podID="ee31dce3-0174-49a2-afce-4cb0e5838256" containerID="5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4" exitCode=0 Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.078882 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" event={"ID":"ee31dce3-0174-49a2-afce-4cb0e5838256","Type":"ContainerDied","Data":"5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4"} Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.078907 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" event={"ID":"ee31dce3-0174-49a2-afce-4cb0e5838256","Type":"ContainerDied","Data":"34daa25d2218ee26a3e47bd23bee95a76e6a6257e0243e9a6db6b53aa296d7f1"} Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.078903 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7745c94d-n92jt" Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.078923 4706 scope.go:117] "RemoveContainer" containerID="5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4" Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.094974 4706 scope.go:117] "RemoveContainer" containerID="5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4" Nov 27 07:14:03 crc kubenswrapper[4706]: E1127 07:14:03.095459 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4\": container with ID starting with 5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4 not found: ID does not exist" containerID="5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4" Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.095514 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4"} err="failed to get container status \"5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4\": rpc error: code = NotFound desc = could not find container \"5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4\": container with ID starting with 5abc67dc0ac1f7083607f29bbd9df66dcb701e9d9a7847d1df3b7f661208b2c4 not found: ID does not exist" Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.103432 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c7745c94d-n92jt"] Nov 27 07:14:03 crc kubenswrapper[4706]: I1127 07:14:03.106808 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6c7745c94d-n92jt"] Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.055622 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85d7648d75-xs548"] Nov 27 07:14:04 crc kubenswrapper[4706]: E1127 07:14:04.055885 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee31dce3-0174-49a2-afce-4cb0e5838256" containerName="controller-manager" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.055901 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee31dce3-0174-49a2-afce-4cb0e5838256" containerName="controller-manager" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.056024 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee31dce3-0174-49a2-afce-4cb0e5838256" containerName="controller-manager" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.056641 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.058625 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.058792 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.059955 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.060105 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.061008 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.061423 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.068537 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85d7648d75-xs548"] Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.078914 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.226285 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b64569d-14ff-4dcd-9368-b51134902796-serving-cert\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.226357 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-config\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.226397 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-client-ca\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.226422 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfs9q\" (UniqueName: \"kubernetes.io/projected/4b64569d-14ff-4dcd-9368-b51134902796-kube-api-access-wfs9q\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.226738 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-proxy-ca-bundles\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.328207 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-proxy-ca-bundles\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.328290 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b64569d-14ff-4dcd-9368-b51134902796-serving-cert\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.328344 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-config\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.328380 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-client-ca\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.328414 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfs9q\" (UniqueName: \"kubernetes.io/projected/4b64569d-14ff-4dcd-9368-b51134902796-kube-api-access-wfs9q\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.329382 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-proxy-ca-bundles\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.329429 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-client-ca\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.329812 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b64569d-14ff-4dcd-9368-b51134902796-config\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.344426 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b64569d-14ff-4dcd-9368-b51134902796-serving-cert\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.348501 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfs9q\" (UniqueName: \"kubernetes.io/projected/4b64569d-14ff-4dcd-9368-b51134902796-kube-api-access-wfs9q\") pod \"controller-manager-85d7648d75-xs548\" (UID: \"4b64569d-14ff-4dcd-9368-b51134902796\") " pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.369961 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.789156 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee31dce3-0174-49a2-afce-4cb0e5838256" path="/var/lib/kubelet/pods/ee31dce3-0174-49a2-afce-4cb0e5838256/volumes" Nov 27 07:14:04 crc kubenswrapper[4706]: I1127 07:14:04.789914 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85d7648d75-xs548"] Nov 27 07:14:05 crc kubenswrapper[4706]: I1127 07:14:05.090855 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" event={"ID":"4b64569d-14ff-4dcd-9368-b51134902796","Type":"ContainerStarted","Data":"6b7ffb176bc36f9b1cd8fba4ffdacb2609cd4e0e396d1374215520bb53b05494"} Nov 27 07:14:05 crc kubenswrapper[4706]: I1127 07:14:05.091262 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:05 crc kubenswrapper[4706]: I1127 07:14:05.091279 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" event={"ID":"4b64569d-14ff-4dcd-9368-b51134902796","Type":"ContainerStarted","Data":"460092e6f46e1bf14363ce7640cbcc6a37bf7475646a9d0944bb35669a417849"} Nov 27 07:14:05 crc kubenswrapper[4706]: I1127 07:14:05.097018 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" Nov 27 07:14:05 crc kubenswrapper[4706]: I1127 07:14:05.112258 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85d7648d75-xs548" podStartSLOduration=3.112218338 podStartE2EDuration="3.112218338s" podCreationTimestamp="2025-11-27 07:14:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:14:05.109775109 +0000 UTC m=+328.999365929" watchObservedRunningTime="2025-11-27 07:14:05.112218338 +0000 UTC m=+329.001809148" Nov 27 07:14:15 crc kubenswrapper[4706]: I1127 07:14:15.177853 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:14:15 crc kubenswrapper[4706]: I1127 07:14:15.178475 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.123060 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6"] Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.123922 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" podUID="8cef8686-b75a-4164-ab52-c3500ed736ef" containerName="route-controller-manager" containerID="cri-o://49ed01abacbfe1fba1a47421ae6456e9062242a96fb7957db1c08aa802d20f00" gracePeriod=30 Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.299008 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cef8686-b75a-4164-ab52-c3500ed736ef" containerID="49ed01abacbfe1fba1a47421ae6456e9062242a96fb7957db1c08aa802d20f00" exitCode=0 Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.299236 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" event={"ID":"8cef8686-b75a-4164-ab52-c3500ed736ef","Type":"ContainerDied","Data":"49ed01abacbfe1fba1a47421ae6456e9062242a96fb7957db1c08aa802d20f00"} Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.501744 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.606185 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jklwg\" (UniqueName: \"kubernetes.io/projected/8cef8686-b75a-4164-ab52-c3500ed736ef-kube-api-access-jklwg\") pod \"8cef8686-b75a-4164-ab52-c3500ed736ef\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.606271 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-config\") pod \"8cef8686-b75a-4164-ab52-c3500ed736ef\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.606292 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cef8686-b75a-4164-ab52-c3500ed736ef-serving-cert\") pod \"8cef8686-b75a-4164-ab52-c3500ed736ef\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.606307 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-client-ca\") pod \"8cef8686-b75a-4164-ab52-c3500ed736ef\" (UID: \"8cef8686-b75a-4164-ab52-c3500ed736ef\") " Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.607083 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-client-ca" (OuterVolumeSpecName: "client-ca") pod "8cef8686-b75a-4164-ab52-c3500ed736ef" (UID: "8cef8686-b75a-4164-ab52-c3500ed736ef"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.607132 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-config" (OuterVolumeSpecName: "config") pod "8cef8686-b75a-4164-ab52-c3500ed736ef" (UID: "8cef8686-b75a-4164-ab52-c3500ed736ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.619692 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cef8686-b75a-4164-ab52-c3500ed736ef-kube-api-access-jklwg" (OuterVolumeSpecName: "kube-api-access-jklwg") pod "8cef8686-b75a-4164-ab52-c3500ed736ef" (UID: "8cef8686-b75a-4164-ab52-c3500ed736ef"). InnerVolumeSpecName "kube-api-access-jklwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.619695 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cef8686-b75a-4164-ab52-c3500ed736ef-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cef8686-b75a-4164-ab52-c3500ed736ef" (UID: "8cef8686-b75a-4164-ab52-c3500ed736ef"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.708207 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jklwg\" (UniqueName: \"kubernetes.io/projected/8cef8686-b75a-4164-ab52-c3500ed736ef-kube-api-access-jklwg\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.708264 4706 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.708275 4706 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cef8686-b75a-4164-ab52-c3500ed736ef-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:42 crc kubenswrapper[4706]: I1127 07:14:42.708288 4706 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cef8686-b75a-4164-ab52-c3500ed736ef-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:14:43 crc kubenswrapper[4706]: I1127 07:14:43.305174 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" event={"ID":"8cef8686-b75a-4164-ab52-c3500ed736ef","Type":"ContainerDied","Data":"0181d14a94d60067d348a54dcf78b61eaf0be42b5054283ea2bd73328d025d84"} Nov 27 07:14:43 crc kubenswrapper[4706]: I1127 07:14:43.305286 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6" Nov 27 07:14:43 crc kubenswrapper[4706]: I1127 07:14:43.305307 4706 scope.go:117] "RemoveContainer" containerID="49ed01abacbfe1fba1a47421ae6456e9062242a96fb7957db1c08aa802d20f00" Nov 27 07:14:43 crc kubenswrapper[4706]: I1127 07:14:43.339524 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6"] Nov 27 07:14:43 crc kubenswrapper[4706]: I1127 07:14:43.339765 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-696c5bc4d6-lsdl6"] Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:43.849683 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ml6l7"] Nov 27 07:14:44 crc kubenswrapper[4706]: E1127 07:14:43.850160 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cef8686-b75a-4164-ab52-c3500ed736ef" containerName="route-controller-manager" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:43.850172 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cef8686-b75a-4164-ab52-c3500ed736ef" containerName="route-controller-manager" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:43.850322 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cef8686-b75a-4164-ab52-c3500ed736ef" containerName="route-controller-manager" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:43.850726 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:43.859774 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ml6l7"] Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.024591 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-registry-tls\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.024711 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.024748 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.024820 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-registry-certificates\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.024892 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56dl6\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-kube-api-access-56dl6\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.024926 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.025119 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-trusted-ca\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.025139 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-bound-sa-token\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.045365 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.079470 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk"] Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.080093 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.082050 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.082191 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.082372 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.082480 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.082641 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.082738 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.090185 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk"] Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.126485 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-bound-sa-token\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.126536 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-registry-tls\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.126564 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.126621 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-registry-certificates\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.126650 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56dl6\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-kube-api-access-56dl6\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.126672 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.126691 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-trusted-ca\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.127810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-trusted-ca\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.129386 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.129585 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-registry-certificates\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.133142 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.133535 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-registry-tls\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.145181 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-bound-sa-token\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.152392 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56dl6\" (UniqueName: \"kubernetes.io/projected/607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9-kube-api-access-56dl6\") pod \"image-registry-66df7c8f76-ml6l7\" (UID: \"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9\") " pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.168160 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.227870 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-client-ca\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.228103 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-kube-api-access-th28m\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.228191 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-serving-cert\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.228297 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-config\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.329362 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-config\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.329694 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-serving-cert\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.329734 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-client-ca\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.329786 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-kube-api-access-th28m\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.330802 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-config\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.330997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-client-ca\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.334884 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-serving-cert\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.346335 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/2dfc474e-77a5-4149-b3a5-b92036d5c9d9-kube-api-access-th28m\") pod \"route-controller-manager-5646b8fc57-rjdlk\" (UID: \"2dfc474e-77a5-4149-b3a5-b92036d5c9d9\") " pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.410190 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.545086 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ml6l7"] Nov 27 07:14:44 crc kubenswrapper[4706]: W1127 07:14:44.555066 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod607c9c07_0ca0_4e11_ae9d_f0cea65ae5a9.slice/crio-faba93bb970fb281497d493dc7d276f8836988743be4eb205cf02a6ecdefe8d7 WatchSource:0}: Error finding container faba93bb970fb281497d493dc7d276f8836988743be4eb205cf02a6ecdefe8d7: Status 404 returned error can't find the container with id faba93bb970fb281497d493dc7d276f8836988743be4eb205cf02a6ecdefe8d7 Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.788267 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cef8686-b75a-4164-ab52-c3500ed736ef" path="/var/lib/kubelet/pods/8cef8686-b75a-4164-ab52-c3500ed736ef/volumes" Nov 27 07:14:44 crc kubenswrapper[4706]: I1127 07:14:44.832585 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk"] Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.178058 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.178427 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.316101 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" event={"ID":"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9","Type":"ContainerStarted","Data":"e610eeb7d23f9c257ec461dad7d35b548e3f284480ea523190da5fc4c9a9b2cc"} Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.316140 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" event={"ID":"607c9c07-0ca0-4e11-ae9d-f0cea65ae5a9","Type":"ContainerStarted","Data":"faba93bb970fb281497d493dc7d276f8836988743be4eb205cf02a6ecdefe8d7"} Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.317124 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.320463 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" event={"ID":"2dfc474e-77a5-4149-b3a5-b92036d5c9d9","Type":"ContainerStarted","Data":"0b193b841cfaa97a0d1be51db4c649cfbb0e4815e633383028c7d093e9598647"} Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.320495 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" event={"ID":"2dfc474e-77a5-4149-b3a5-b92036d5c9d9","Type":"ContainerStarted","Data":"9ce1450a0499c30bc320089763c13cc46cbf89f4735fe3f81ac0b6ffca033fc6"} Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.321117 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.325951 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.341520 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" podStartSLOduration=2.34150562 podStartE2EDuration="2.34150562s" podCreationTimestamp="2025-11-27 07:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:14:45.336755246 +0000 UTC m=+369.226346066" watchObservedRunningTime="2025-11-27 07:14:45.34150562 +0000 UTC m=+369.231096430" Nov 27 07:14:45 crc kubenswrapper[4706]: I1127 07:14:45.359577 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5646b8fc57-rjdlk" podStartSLOduration=3.359561557 podStartE2EDuration="3.359561557s" podCreationTimestamp="2025-11-27 07:14:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:14:45.357570722 +0000 UTC m=+369.247161532" watchObservedRunningTime="2025-11-27 07:14:45.359561557 +0000 UTC m=+369.249152367" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.149149 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s"] Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.151741 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.155064 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.156406 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.170516 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s"] Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.232515 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dcbae6f3-ee39-47d9-8b8c-15f639da6519-secret-volume\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.232578 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcbae6f3-ee39-47d9-8b8c-15f639da6519-config-volume\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.232657 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9fvs\" (UniqueName: \"kubernetes.io/projected/dcbae6f3-ee39-47d9-8b8c-15f639da6519-kube-api-access-h9fvs\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.334083 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9fvs\" (UniqueName: \"kubernetes.io/projected/dcbae6f3-ee39-47d9-8b8c-15f639da6519-kube-api-access-h9fvs\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.334180 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dcbae6f3-ee39-47d9-8b8c-15f639da6519-secret-volume\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.334890 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcbae6f3-ee39-47d9-8b8c-15f639da6519-config-volume\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.335176 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcbae6f3-ee39-47d9-8b8c-15f639da6519-config-volume\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.341568 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dcbae6f3-ee39-47d9-8b8c-15f639da6519-secret-volume\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.348883 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9fvs\" (UniqueName: \"kubernetes.io/projected/dcbae6f3-ee39-47d9-8b8c-15f639da6519-kube-api-access-h9fvs\") pod \"collect-profiles-29403795-2x75s\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.516236 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:00 crc kubenswrapper[4706]: I1127 07:15:00.703729 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s"] Nov 27 07:15:01 crc kubenswrapper[4706]: I1127 07:15:01.420013 4706 generic.go:334] "Generic (PLEG): container finished" podID="dcbae6f3-ee39-47d9-8b8c-15f639da6519" containerID="370a1668203fc0b7e6ae4383bbdb1823ea5bd5505875e03f79dc25d8a7b173d6" exitCode=0 Nov 27 07:15:01 crc kubenswrapper[4706]: I1127 07:15:01.420135 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" event={"ID":"dcbae6f3-ee39-47d9-8b8c-15f639da6519","Type":"ContainerDied","Data":"370a1668203fc0b7e6ae4383bbdb1823ea5bd5505875e03f79dc25d8a7b173d6"} Nov 27 07:15:01 crc kubenswrapper[4706]: I1127 07:15:01.420424 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" event={"ID":"dcbae6f3-ee39-47d9-8b8c-15f639da6519","Type":"ContainerStarted","Data":"9f994d61e8e8444c57dfa485b2c482f236491aed64c07510208db941f2422065"} Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.743634 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.868531 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dcbae6f3-ee39-47d9-8b8c-15f639da6519-secret-volume\") pod \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.868587 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcbae6f3-ee39-47d9-8b8c-15f639da6519-config-volume\") pod \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.868608 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9fvs\" (UniqueName: \"kubernetes.io/projected/dcbae6f3-ee39-47d9-8b8c-15f639da6519-kube-api-access-h9fvs\") pod \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\" (UID: \"dcbae6f3-ee39-47d9-8b8c-15f639da6519\") " Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.869628 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcbae6f3-ee39-47d9-8b8c-15f639da6519-config-volume" (OuterVolumeSpecName: "config-volume") pod "dcbae6f3-ee39-47d9-8b8c-15f639da6519" (UID: "dcbae6f3-ee39-47d9-8b8c-15f639da6519"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.875189 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcbae6f3-ee39-47d9-8b8c-15f639da6519-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dcbae6f3-ee39-47d9-8b8c-15f639da6519" (UID: "dcbae6f3-ee39-47d9-8b8c-15f639da6519"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.876375 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcbae6f3-ee39-47d9-8b8c-15f639da6519-kube-api-access-h9fvs" (OuterVolumeSpecName: "kube-api-access-h9fvs") pod "dcbae6f3-ee39-47d9-8b8c-15f639da6519" (UID: "dcbae6f3-ee39-47d9-8b8c-15f639da6519"). InnerVolumeSpecName "kube-api-access-h9fvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.970211 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dcbae6f3-ee39-47d9-8b8c-15f639da6519-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.970263 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcbae6f3-ee39-47d9-8b8c-15f639da6519-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:02 crc kubenswrapper[4706]: I1127 07:15:02.970281 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9fvs\" (UniqueName: \"kubernetes.io/projected/dcbae6f3-ee39-47d9-8b8c-15f639da6519-kube-api-access-h9fvs\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:03 crc kubenswrapper[4706]: I1127 07:15:03.432484 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" event={"ID":"dcbae6f3-ee39-47d9-8b8c-15f639da6519","Type":"ContainerDied","Data":"9f994d61e8e8444c57dfa485b2c482f236491aed64c07510208db941f2422065"} Nov 27 07:15:03 crc kubenswrapper[4706]: I1127 07:15:03.432521 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f994d61e8e8444c57dfa485b2c482f236491aed64c07510208db941f2422065" Nov 27 07:15:03 crc kubenswrapper[4706]: I1127 07:15:03.432544 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403795-2x75s" Nov 27 07:15:04 crc kubenswrapper[4706]: I1127 07:15:04.172928 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-ml6l7" Nov 27 07:15:04 crc kubenswrapper[4706]: I1127 07:15:04.236925 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2gpf6"] Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.178055 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.178639 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.178688 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.179416 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8bcee89027fe370848c3797738380643eda24e7a750e5b718af96d44bf36e027"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.179479 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://8bcee89027fe370848c3797738380643eda24e7a750e5b718af96d44bf36e027" gracePeriod=600 Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.517550 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="8bcee89027fe370848c3797738380643eda24e7a750e5b718af96d44bf36e027" exitCode=0 Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.517628 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"8bcee89027fe370848c3797738380643eda24e7a750e5b718af96d44bf36e027"} Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.517923 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"7171128fb93f11a5548129b54f5af0ff478bef072ea546bb401cec5c726024db"} Nov 27 07:15:15 crc kubenswrapper[4706]: I1127 07:15:15.517943 4706 scope.go:117] "RemoveContainer" containerID="dd7fb877f6182061748365f24158ef4b24809b501b184951ed8bad250d2f048f" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.281422 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" podUID="bdec2b7f-560f-44e9-b665-2ff892495d19" containerName="registry" containerID="cri-o://5918a2b15e0e21b34613f46b8f0cdabf49ad15328d82288b5c58349bbe7b2729" gracePeriod=30 Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.611118 4706 generic.go:334] "Generic (PLEG): container finished" podID="bdec2b7f-560f-44e9-b665-2ff892495d19" containerID="5918a2b15e0e21b34613f46b8f0cdabf49ad15328d82288b5c58349bbe7b2729" exitCode=0 Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.611200 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" event={"ID":"bdec2b7f-560f-44e9-b665-2ff892495d19","Type":"ContainerDied","Data":"5918a2b15e0e21b34613f46b8f0cdabf49ad15328d82288b5c58349bbe7b2729"} Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.678669 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.788861 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-bound-sa-token\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.788919 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6fdj\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-kube-api-access-r6fdj\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.789182 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.789248 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bdec2b7f-560f-44e9-b665-2ff892495d19-ca-trust-extracted\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.789281 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bdec2b7f-560f-44e9-b665-2ff892495d19-installation-pull-secrets\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.789334 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-certificates\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.789377 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-trusted-ca\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.789471 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-tls\") pod \"bdec2b7f-560f-44e9-b665-2ff892495d19\" (UID: \"bdec2b7f-560f-44e9-b665-2ff892495d19\") " Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.790739 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.790990 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.795795 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.796258 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-kube-api-access-r6fdj" (OuterVolumeSpecName: "kube-api-access-r6fdj") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "kube-api-access-r6fdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.796255 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdec2b7f-560f-44e9-b665-2ff892495d19-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.798512 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.803855 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.817073 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdec2b7f-560f-44e9-b665-2ff892495d19-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bdec2b7f-560f-44e9-b665-2ff892495d19" (UID: "bdec2b7f-560f-44e9-b665-2ff892495d19"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.891184 4706 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.891298 4706 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.891323 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6fdj\" (UniqueName: \"kubernetes.io/projected/bdec2b7f-560f-44e9-b665-2ff892495d19-kube-api-access-r6fdj\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.891343 4706 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bdec2b7f-560f-44e9-b665-2ff892495d19-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.891362 4706 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bdec2b7f-560f-44e9-b665-2ff892495d19-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.891380 4706 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:29 crc kubenswrapper[4706]: I1127 07:15:29.891398 4706 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdec2b7f-560f-44e9-b665-2ff892495d19-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 07:15:30 crc kubenswrapper[4706]: I1127 07:15:30.619985 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" event={"ID":"bdec2b7f-560f-44e9-b665-2ff892495d19","Type":"ContainerDied","Data":"163909d20be1a0aa6de660b3c2484691dc632973bb1e9234234997dacb42a10e"} Nov 27 07:15:30 crc kubenswrapper[4706]: I1127 07:15:30.620082 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2gpf6" Nov 27 07:15:30 crc kubenswrapper[4706]: I1127 07:15:30.620090 4706 scope.go:117] "RemoveContainer" containerID="5918a2b15e0e21b34613f46b8f0cdabf49ad15328d82288b5c58349bbe7b2729" Nov 27 07:15:30 crc kubenswrapper[4706]: I1127 07:15:30.672726 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2gpf6"] Nov 27 07:15:30 crc kubenswrapper[4706]: I1127 07:15:30.680459 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2gpf6"] Nov 27 07:15:30 crc kubenswrapper[4706]: I1127 07:15:30.784427 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdec2b7f-560f-44e9-b665-2ff892495d19" path="/var/lib/kubelet/pods/bdec2b7f-560f-44e9-b665-2ff892495d19/volumes" Nov 27 07:17:15 crc kubenswrapper[4706]: I1127 07:17:15.178349 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:17:15 crc kubenswrapper[4706]: I1127 07:17:15.179045 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:17:45 crc kubenswrapper[4706]: I1127 07:17:45.177877 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:17:45 crc kubenswrapper[4706]: I1127 07:17:45.178658 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.178544 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.179175 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.179295 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.180151 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7171128fb93f11a5548129b54f5af0ff478bef072ea546bb401cec5c726024db"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.180287 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://7171128fb93f11a5548129b54f5af0ff478bef072ea546bb401cec5c726024db" gracePeriod=600 Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.613549 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="7171128fb93f11a5548129b54f5af0ff478bef072ea546bb401cec5c726024db" exitCode=0 Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.613589 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"7171128fb93f11a5548129b54f5af0ff478bef072ea546bb401cec5c726024db"} Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.614188 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"4cee0d2f37ffae96c2927fba888af9c607b55ddd51e1dd0044435dc5bccfed27"} Nov 27 07:18:15 crc kubenswrapper[4706]: I1127 07:18:15.614262 4706 scope.go:117] "RemoveContainer" containerID="8bcee89027fe370848c3797738380643eda24e7a750e5b718af96d44bf36e027" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.307368 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhhb"] Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.308926 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-controller" containerID="cri-o://f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.309036 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="nbdb" containerID="cri-o://93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.309115 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="northd" containerID="cri-o://575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.309203 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.309369 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-node" containerID="cri-o://c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.309448 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-acl-logging" containerID="cri-o://20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.309653 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="sbdb" containerID="cri-o://cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.360218 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" containerID="cri-o://12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" gracePeriod=30 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.643145 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/3.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.645356 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovn-acl-logging/0.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.645789 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovn-controller/0.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.646209 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.700623 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-697w8"] Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.700892 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="nbdb" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.700917 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="nbdb" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.700933 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.700942 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.700951 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kubecfg-setup" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.700960 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kubecfg-setup" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.700973 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-node" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.700981 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-node" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.700991 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701000 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701009 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="sbdb" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701017 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="sbdb" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701026 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701033 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701043 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701051 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701062 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-acl-logging" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701069 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-acl-logging" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701080 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701088 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701099 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbae6f3-ee39-47d9-8b8c-15f639da6519" containerName="collect-profiles" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701107 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbae6f3-ee39-47d9-8b8c-15f639da6519" containerName="collect-profiles" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701119 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdec2b7f-560f-44e9-b665-2ff892495d19" containerName="registry" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701126 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdec2b7f-560f-44e9-b665-2ff892495d19" containerName="registry" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701138 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="northd" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701145 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="northd" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701157 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701165 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701297 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701313 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701327 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="sbdb" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701335 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701345 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="nbdb" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701355 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-node" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701363 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701373 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701383 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-acl-logging" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701394 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701404 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdec2b7f-560f-44e9-b665-2ff892495d19" containerName="registry" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701416 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="northd" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701425 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovn-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701435 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbae6f3-ee39-47d9-8b8c-15f639da6519" containerName="collect-profiles" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.701551 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.701564 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerName="ovnkube-controller" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.703725 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.710297 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/2.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.711288 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/1.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.711337 4706 generic.go:334] "Generic (PLEG): container finished" podID="0e33550a-5308-4e22-9c81-8a12e97f44ef" containerID="408dbfa8d92fabe6661bc2452e5738f59d4453f29fd4329bb8b6037bc5acd519" exitCode=2 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.711403 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerDied","Data":"408dbfa8d92fabe6661bc2452e5738f59d4453f29fd4329bb8b6037bc5acd519"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.711440 4706 scope.go:117] "RemoveContainer" containerID="5fe647f5147e63e0a3c9295afe8e5d564ff58d9fd8ea192debdf7831041ba3b6" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.712091 4706 scope.go:117] "RemoveContainer" containerID="408dbfa8d92fabe6661bc2452e5738f59d4453f29fd4329bb8b6037bc5acd519" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.712352 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m645h_openshift-multus(0e33550a-5308-4e22-9c81-8a12e97f44ef)\"" pod="openshift-multus/multus-m645h" podUID="0e33550a-5308-4e22-9c81-8a12e97f44ef" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.715634 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovnkube-controller/3.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.718289 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovn-acl-logging/0.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719046 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhhb_d1b93703-3a71-49b6-bff6-b4d314006ddd/ovn-controller/0.log" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719770 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" exitCode=0 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719794 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" exitCode=0 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719803 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" exitCode=0 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719812 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" exitCode=0 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719819 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" exitCode=0 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719825 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" exitCode=0 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719832 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" exitCode=143 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719838 4706 generic.go:334] "Generic (PLEG): container finished" podID="d1b93703-3a71-49b6-bff6-b4d314006ddd" containerID="f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" exitCode=143 Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719853 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719872 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719883 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719893 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719903 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719912 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719921 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719930 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719936 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719941 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719947 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719953 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719959 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719965 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719971 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719977 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719984 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719991 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.719997 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720002 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720007 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720012 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720017 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720022 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720027 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720032 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720037 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720044 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720051 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720057 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720062 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720068 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720073 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720078 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720084 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720089 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720095 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720100 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720107 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" event={"ID":"d1b93703-3a71-49b6-bff6-b4d314006ddd","Type":"ContainerDied","Data":"855a8b79edc691d8ad0e3d12d589287fd4cca9ede85f6439ef7b744883fabcd9"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720114 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720120 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720125 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720130 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720136 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720141 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720165 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720172 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720177 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720182 4706 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.720276 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhhb" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.739464 4706 scope.go:117] "RemoveContainer" containerID="12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.766199 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.787925 4706 scope.go:117] "RemoveContainer" containerID="cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.807343 4706 scope.go:117] "RemoveContainer" containerID="93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.816672 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-ovn\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.816796 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.816880 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-var-lib-openvswitch\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.816947 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.816985 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-env-overrides\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817006 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-netd\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817112 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817457 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovn-node-metrics-cert\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817515 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-openvswitch\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817710 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817766 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-node-log\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817802 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-log-socket\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817819 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-slash\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817848 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-node-log" (OuterVolumeSpecName: "node-log") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817897 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-ovn-kubernetes\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817864 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-log-socket" (OuterVolumeSpecName: "log-socket") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817914 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-bin\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817934 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-slash" (OuterVolumeSpecName: "host-slash") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817952 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-systemd-units\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817961 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817973 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-script-lib\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817986 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818006 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s94lx\" (UniqueName: \"kubernetes.io/projected/d1b93703-3a71-49b6-bff6-b4d314006ddd-kube-api-access-s94lx\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.817996 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818039 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818101 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818118 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-netns\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818134 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-systemd\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818155 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-config\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818171 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-etc-openvswitch\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818186 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-kubelet\") pod \"d1b93703-3a71-49b6-bff6-b4d314006ddd\" (UID: \"d1b93703-3a71-49b6-bff6-b4d314006ddd\") " Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818251 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818264 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818327 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovn-node-metrics-cert\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818355 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-systemd\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818367 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818438 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-cni-netd\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818480 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-env-overrides\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818562 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-log-socket\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818647 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818694 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-kubelet\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818788 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdczq\" (UniqueName: \"kubernetes.io/projected/3fe1c217-ffdd-4531-8337-4d209ebc1a36-kube-api-access-bdczq\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818830 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818924 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-slash\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.818963 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-cni-bin\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819004 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovnkube-config\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819034 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-node-log\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819058 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-run-ovn-kubernetes\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819103 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-ovn\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819149 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-var-lib-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819176 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-etc-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819317 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-systemd-units\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819378 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovnkube-script-lib\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819401 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-run-netns\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819498 4706 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819520 4706 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819536 4706 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819552 4706 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819564 4706 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819616 4706 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819627 4706 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819636 4706 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819647 4706 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819660 4706 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819670 4706 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819681 4706 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819694 4706 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-node-log\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819703 4706 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-log-socket\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819710 4706 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-host-slash\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.819894 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.820905 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.823275 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.823449 4706 scope.go:117] "RemoveContainer" containerID="575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.823472 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b93703-3a71-49b6-bff6-b4d314006ddd-kube-api-access-s94lx" (OuterVolumeSpecName: "kube-api-access-s94lx") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "kube-api-access-s94lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.830598 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d1b93703-3a71-49b6-bff6-b4d314006ddd" (UID: "d1b93703-3a71-49b6-bff6-b4d314006ddd"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.835213 4706 scope.go:117] "RemoveContainer" containerID="738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.845487 4706 scope.go:117] "RemoveContainer" containerID="c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.862287 4706 scope.go:117] "RemoveContainer" containerID="20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.873942 4706 scope.go:117] "RemoveContainer" containerID="f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.884467 4706 scope.go:117] "RemoveContainer" containerID="839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.895915 4706 scope.go:117] "RemoveContainer" containerID="12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.896287 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": container with ID starting with 12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c not found: ID does not exist" containerID="12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.896337 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} err="failed to get container status \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": rpc error: code = NotFound desc = could not find container \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": container with ID starting with 12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.896359 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.896605 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": container with ID starting with b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c not found: ID does not exist" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.896636 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} err="failed to get container status \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": rpc error: code = NotFound desc = could not find container \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": container with ID starting with b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.896660 4706 scope.go:117] "RemoveContainer" containerID="cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.896882 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": container with ID starting with cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194 not found: ID does not exist" containerID="cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.896920 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} err="failed to get container status \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": rpc error: code = NotFound desc = could not find container \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": container with ID starting with cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.896940 4706 scope.go:117] "RemoveContainer" containerID="93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.897196 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": container with ID starting with 93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515 not found: ID does not exist" containerID="93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.897272 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} err="failed to get container status \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": rpc error: code = NotFound desc = could not find container \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": container with ID starting with 93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.897316 4706 scope.go:117] "RemoveContainer" containerID="575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.897577 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": container with ID starting with 575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a not found: ID does not exist" containerID="575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.897614 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} err="failed to get container status \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": rpc error: code = NotFound desc = could not find container \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": container with ID starting with 575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.897627 4706 scope.go:117] "RemoveContainer" containerID="738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.897812 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": container with ID starting with 738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a not found: ID does not exist" containerID="738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.897841 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} err="failed to get container status \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": rpc error: code = NotFound desc = could not find container \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": container with ID starting with 738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.897857 4706 scope.go:117] "RemoveContainer" containerID="c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.898044 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": container with ID starting with c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8 not found: ID does not exist" containerID="c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898064 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} err="failed to get container status \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": rpc error: code = NotFound desc = could not find container \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": container with ID starting with c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898076 4706 scope.go:117] "RemoveContainer" containerID="20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.898329 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": container with ID starting with 20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8 not found: ID does not exist" containerID="20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898359 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} err="failed to get container status \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": rpc error: code = NotFound desc = could not find container \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": container with ID starting with 20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898379 4706 scope.go:117] "RemoveContainer" containerID="f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.898591 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": container with ID starting with f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9 not found: ID does not exist" containerID="f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898613 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} err="failed to get container status \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": rpc error: code = NotFound desc = could not find container \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": container with ID starting with f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898627 4706 scope.go:117] "RemoveContainer" containerID="839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172" Nov 27 07:18:24 crc kubenswrapper[4706]: E1127 07:18:24.898779 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": container with ID starting with 839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172 not found: ID does not exist" containerID="839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898800 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} err="failed to get container status \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": rpc error: code = NotFound desc = could not find container \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": container with ID starting with 839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.898816 4706 scope.go:117] "RemoveContainer" containerID="12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899007 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} err="failed to get container status \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": rpc error: code = NotFound desc = could not find container \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": container with ID starting with 12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899044 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899234 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} err="failed to get container status \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": rpc error: code = NotFound desc = could not find container \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": container with ID starting with b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899255 4706 scope.go:117] "RemoveContainer" containerID="cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899551 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} err="failed to get container status \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": rpc error: code = NotFound desc = could not find container \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": container with ID starting with cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899598 4706 scope.go:117] "RemoveContainer" containerID="93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899785 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} err="failed to get container status \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": rpc error: code = NotFound desc = could not find container \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": container with ID starting with 93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899809 4706 scope.go:117] "RemoveContainer" containerID="575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.899984 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} err="failed to get container status \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": rpc error: code = NotFound desc = could not find container \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": container with ID starting with 575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900003 4706 scope.go:117] "RemoveContainer" containerID="738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900203 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} err="failed to get container status \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": rpc error: code = NotFound desc = could not find container \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": container with ID starting with 738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900258 4706 scope.go:117] "RemoveContainer" containerID="c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900499 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} err="failed to get container status \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": rpc error: code = NotFound desc = could not find container \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": container with ID starting with c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900519 4706 scope.go:117] "RemoveContainer" containerID="20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900719 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} err="failed to get container status \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": rpc error: code = NotFound desc = could not find container \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": container with ID starting with 20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900736 4706 scope.go:117] "RemoveContainer" containerID="f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900940 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} err="failed to get container status \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": rpc error: code = NotFound desc = could not find container \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": container with ID starting with f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.900967 4706 scope.go:117] "RemoveContainer" containerID="839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901159 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} err="failed to get container status \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": rpc error: code = NotFound desc = could not find container \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": container with ID starting with 839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901180 4706 scope.go:117] "RemoveContainer" containerID="12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901422 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} err="failed to get container status \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": rpc error: code = NotFound desc = could not find container \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": container with ID starting with 12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901450 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901660 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} err="failed to get container status \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": rpc error: code = NotFound desc = could not find container \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": container with ID starting with b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901693 4706 scope.go:117] "RemoveContainer" containerID="cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901869 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} err="failed to get container status \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": rpc error: code = NotFound desc = could not find container \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": container with ID starting with cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.901886 4706 scope.go:117] "RemoveContainer" containerID="93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.902109 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} err="failed to get container status \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": rpc error: code = NotFound desc = could not find container \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": container with ID starting with 93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.902976 4706 scope.go:117] "RemoveContainer" containerID="575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.903296 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} err="failed to get container status \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": rpc error: code = NotFound desc = could not find container \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": container with ID starting with 575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.903335 4706 scope.go:117] "RemoveContainer" containerID="738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.903713 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} err="failed to get container status \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": rpc error: code = NotFound desc = could not find container \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": container with ID starting with 738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.903751 4706 scope.go:117] "RemoveContainer" containerID="c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.903991 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} err="failed to get container status \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": rpc error: code = NotFound desc = could not find container \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": container with ID starting with c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.904034 4706 scope.go:117] "RemoveContainer" containerID="20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.904275 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} err="failed to get container status \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": rpc error: code = NotFound desc = could not find container \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": container with ID starting with 20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.904294 4706 scope.go:117] "RemoveContainer" containerID="f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.904524 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} err="failed to get container status \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": rpc error: code = NotFound desc = could not find container \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": container with ID starting with f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.904569 4706 scope.go:117] "RemoveContainer" containerID="839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.904799 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} err="failed to get container status \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": rpc error: code = NotFound desc = could not find container \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": container with ID starting with 839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.904824 4706 scope.go:117] "RemoveContainer" containerID="12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905044 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c"} err="failed to get container status \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": rpc error: code = NotFound desc = could not find container \"12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c\": container with ID starting with 12e6fdb9ae3b84d5f133000eb8e6005994cf8451bb4140b680aa3fcb7892729c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905085 4706 scope.go:117] "RemoveContainer" containerID="b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905339 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c"} err="failed to get container status \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": rpc error: code = NotFound desc = could not find container \"b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c\": container with ID starting with b5b366ae9b3149e83c18400f8a7e77152598abd8339a7cf689abd2a24b90d23c not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905369 4706 scope.go:117] "RemoveContainer" containerID="cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905554 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194"} err="failed to get container status \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": rpc error: code = NotFound desc = could not find container \"cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194\": container with ID starting with cec935502f3d95898a500557cf5bee812a0e0ef2d1b1755f4ce78023c3bdf194 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905574 4706 scope.go:117] "RemoveContainer" containerID="93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905763 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515"} err="failed to get container status \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": rpc error: code = NotFound desc = could not find container \"93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515\": container with ID starting with 93d3d798fc5d6eab345cdfa862c429f3151894a05e0dd9a3288c5804e4b0a515 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905785 4706 scope.go:117] "RemoveContainer" containerID="575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.905984 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a"} err="failed to get container status \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": rpc error: code = NotFound desc = could not find container \"575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a\": container with ID starting with 575c47f262a36329bd475c9a56d0fd214e515e82055562ac6852a91c380bd77a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906009 4706 scope.go:117] "RemoveContainer" containerID="738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906177 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a"} err="failed to get container status \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": rpc error: code = NotFound desc = could not find container \"738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a\": container with ID starting with 738893c7c7db51d5472ee3623e41b88d832712b1317c3a0c67a462a2e4a9096a not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906204 4706 scope.go:117] "RemoveContainer" containerID="c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906453 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8"} err="failed to get container status \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": rpc error: code = NotFound desc = could not find container \"c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8\": container with ID starting with c02bd60e1434aa5c3c1f75998d79dca1c801dde9100ba5470564cbbc8d3d26c8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906478 4706 scope.go:117] "RemoveContainer" containerID="20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906657 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8"} err="failed to get container status \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": rpc error: code = NotFound desc = could not find container \"20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8\": container with ID starting with 20368071c40ded94dc432cabb037474132474b4427ad94cca0305fc4fdc03cb8 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906680 4706 scope.go:117] "RemoveContainer" containerID="f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906851 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9"} err="failed to get container status \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": rpc error: code = NotFound desc = could not find container \"f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9\": container with ID starting with f1a59afe07c0417fbc55d9f3d1968103388d6c14c84c1291f0e748e6e61c67f9 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.906870 4706 scope.go:117] "RemoveContainer" containerID="839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.907061 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172"} err="failed to get container status \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": rpc error: code = NotFound desc = could not find container \"839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172\": container with ID starting with 839cf2e546cce47938e83cd2d7fd439f1b9bb0912db4fe44024ef316f24e2172 not found: ID does not exist" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921244 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921286 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-kubelet\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921324 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdczq\" (UniqueName: \"kubernetes.io/projected/3fe1c217-ffdd-4531-8337-4d209ebc1a36-kube-api-access-bdczq\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921355 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921360 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921363 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-kubelet\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921374 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-slash\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921518 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-cni-bin\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921566 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovnkube-config\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921575 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-cni-bin\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921422 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-slash\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921407 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921624 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-node-log\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921645 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-run-ovn-kubernetes\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921683 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-node-log\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921738 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-ovn\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921772 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-var-lib-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921791 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-etc-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921799 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-run-ovn-kubernetes\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921832 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-ovn\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921856 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-var-lib-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-systemd-units\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921942 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-etc-openvswitch\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921937 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-systemd-units\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.921989 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovnkube-script-lib\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922005 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-run-netns\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovn-node-metrics-cert\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922059 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-systemd\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922097 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-cni-netd\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-env-overrides\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922142 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-log-socket\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922322 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-run-netns\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922366 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922759 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-run-systemd\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922780 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-env-overrides\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922860 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovnkube-config\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922951 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s94lx\" (UniqueName: \"kubernetes.io/projected/d1b93703-3a71-49b6-bff6-b4d314006ddd-kube-api-access-s94lx\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922998 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-host-cni-netd\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.922987 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3fe1c217-ffdd-4531-8337-4d209ebc1a36-log-socket\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.923144 4706 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1b93703-3a71-49b6-bff6-b4d314006ddd-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.923144 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovnkube-script-lib\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.923214 4706 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.923245 4706 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1b93703-3a71-49b6-bff6-b4d314006ddd-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.929959 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3fe1c217-ffdd-4531-8337-4d209ebc1a36-ovn-node-metrics-cert\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:24 crc kubenswrapper[4706]: I1127 07:18:24.936676 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdczq\" (UniqueName: \"kubernetes.io/projected/3fe1c217-ffdd-4531-8337-4d209ebc1a36-kube-api-access-bdczq\") pod \"ovnkube-node-697w8\" (UID: \"3fe1c217-ffdd-4531-8337-4d209ebc1a36\") " pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:25 crc kubenswrapper[4706]: I1127 07:18:25.015953 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:25 crc kubenswrapper[4706]: I1127 07:18:25.047304 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhhb"] Nov 27 07:18:25 crc kubenswrapper[4706]: I1127 07:18:25.054832 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhhb"] Nov 27 07:18:25 crc kubenswrapper[4706]: I1127 07:18:25.725853 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/2.log" Nov 27 07:18:25 crc kubenswrapper[4706]: I1127 07:18:25.728555 4706 generic.go:334] "Generic (PLEG): container finished" podID="3fe1c217-ffdd-4531-8337-4d209ebc1a36" containerID="0c741a5e5f5b7001d5d11e15899e047289181685414f88a9a948f139eb330aee" exitCode=0 Nov 27 07:18:25 crc kubenswrapper[4706]: I1127 07:18:25.728599 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerDied","Data":"0c741a5e5f5b7001d5d11e15899e047289181685414f88a9a948f139eb330aee"} Nov 27 07:18:25 crc kubenswrapper[4706]: I1127 07:18:25.728631 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"5abef326d4edcd2af0946d3962cd2a10261367f53e8f6cc6fd3695f93866b7eb"} Nov 27 07:18:26 crc kubenswrapper[4706]: I1127 07:18:26.738003 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"f6a79f23792dadedd3f6a5d998766c371a4186a97ba38f28b0a3b5e389910cf0"} Nov 27 07:18:26 crc kubenswrapper[4706]: I1127 07:18:26.738575 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"f9c7370f36c84dd0d8b69bf5e61f4654ef35d89ab240b27426e5dadeaaaf90e4"} Nov 27 07:18:26 crc kubenswrapper[4706]: I1127 07:18:26.738591 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"23b96e1d4e4a7de8ce3a03b7c8226c64272db92a1bf22f7110a59f0f269d5a9a"} Nov 27 07:18:26 crc kubenswrapper[4706]: I1127 07:18:26.738602 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"0153e2a032566e821203c42390003ada4a8f7bc0ad68af04cdbc7d66147059c0"} Nov 27 07:18:26 crc kubenswrapper[4706]: I1127 07:18:26.738612 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"d7d71d8e8da8ac77c34fb4adfd467874842a071df1001da52516a49bafcdbd2a"} Nov 27 07:18:26 crc kubenswrapper[4706]: I1127 07:18:26.738623 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"9fc88d6372bdef9e76028a9fe998f28bf0211e4d6300a235ff3ca2f4547406e2"} Nov 27 07:18:26 crc kubenswrapper[4706]: I1127 07:18:26.784333 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b93703-3a71-49b6-bff6-b4d314006ddd" path="/var/lib/kubelet/pods/d1b93703-3a71-49b6-bff6-b4d314006ddd/volumes" Nov 27 07:18:28 crc kubenswrapper[4706]: I1127 07:18:28.772739 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"1dd599f51b964ee0050060bc2831b29770cbf5b91332d90765d63b3cc669a020"} Nov 27 07:18:31 crc kubenswrapper[4706]: I1127 07:18:31.800299 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" event={"ID":"3fe1c217-ffdd-4531-8337-4d209ebc1a36","Type":"ContainerStarted","Data":"3037d80c0ff33e034eda296f4f61cc8806ebca353dd94edad155cf6bd9195640"} Nov 27 07:18:31 crc kubenswrapper[4706]: I1127 07:18:31.800667 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:31 crc kubenswrapper[4706]: I1127 07:18:31.800682 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:31 crc kubenswrapper[4706]: I1127 07:18:31.800694 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:31 crc kubenswrapper[4706]: I1127 07:18:31.842814 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" podStartSLOduration=7.842784622 podStartE2EDuration="7.842784622s" podCreationTimestamp="2025-11-27 07:18:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:18:31.841197934 +0000 UTC m=+595.730788764" watchObservedRunningTime="2025-11-27 07:18:31.842784622 +0000 UTC m=+595.732375472" Nov 27 07:18:31 crc kubenswrapper[4706]: I1127 07:18:31.844628 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:31 crc kubenswrapper[4706]: I1127 07:18:31.847560 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:34 crc kubenswrapper[4706]: I1127 07:18:34.777156 4706 scope.go:117] "RemoveContainer" containerID="408dbfa8d92fabe6661bc2452e5738f59d4453f29fd4329bb8b6037bc5acd519" Nov 27 07:18:34 crc kubenswrapper[4706]: E1127 07:18:34.777870 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m645h_openshift-multus(0e33550a-5308-4e22-9c81-8a12e97f44ef)\"" pod="openshift-multus/multus-m645h" podUID="0e33550a-5308-4e22-9c81-8a12e97f44ef" Nov 27 07:18:45 crc kubenswrapper[4706]: I1127 07:18:45.776486 4706 scope.go:117] "RemoveContainer" containerID="408dbfa8d92fabe6661bc2452e5738f59d4453f29fd4329bb8b6037bc5acd519" Nov 27 07:18:46 crc kubenswrapper[4706]: I1127 07:18:46.882710 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m645h_0e33550a-5308-4e22-9c81-8a12e97f44ef/kube-multus/2.log" Nov 27 07:18:46 crc kubenswrapper[4706]: I1127 07:18:46.883015 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m645h" event={"ID":"0e33550a-5308-4e22-9c81-8a12e97f44ef","Type":"ContainerStarted","Data":"26cc995d53fc322b0131189308a14eccd9a0e75a2000401649421293339e5d10"} Nov 27 07:18:55 crc kubenswrapper[4706]: I1127 07:18:55.046004 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-697w8" Nov 27 07:18:55 crc kubenswrapper[4706]: I1127 07:18:55.936899 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv"] Nov 27 07:18:55 crc kubenswrapper[4706]: I1127 07:18:55.937930 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:55 crc kubenswrapper[4706]: I1127 07:18:55.939392 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 07:18:55 crc kubenswrapper[4706]: I1127 07:18:55.954802 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv"] Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.047271 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.047357 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.047447 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw8ml\" (UniqueName: \"kubernetes.io/projected/19aee991-c9f9-4d97-9af1-3162462f7d1c-kube-api-access-lw8ml\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.148027 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw8ml\" (UniqueName: \"kubernetes.io/projected/19aee991-c9f9-4d97-9af1-3162462f7d1c-kube-api-access-lw8ml\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.148105 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.148161 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.148705 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.148806 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.172589 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw8ml\" (UniqueName: \"kubernetes.io/projected/19aee991-c9f9-4d97-9af1-3162462f7d1c-kube-api-access-lw8ml\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.273149 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.592604 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv"] Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.941438 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" event={"ID":"19aee991-c9f9-4d97-9af1-3162462f7d1c","Type":"ContainerStarted","Data":"f4af92f2b0e99f6e05936eeefe70f3496c19f7f4a87e3b123f7ca89258476c0a"} Nov 27 07:18:56 crc kubenswrapper[4706]: I1127 07:18:56.942104 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" event={"ID":"19aee991-c9f9-4d97-9af1-3162462f7d1c","Type":"ContainerStarted","Data":"7e915de1cdfff0ef18179e5760beed1c63e3cbf7c5de5111a9f3876794a1c9a5"} Nov 27 07:18:57 crc kubenswrapper[4706]: I1127 07:18:57.949802 4706 generic.go:334] "Generic (PLEG): container finished" podID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerID="f4af92f2b0e99f6e05936eeefe70f3496c19f7f4a87e3b123f7ca89258476c0a" exitCode=0 Nov 27 07:18:57 crc kubenswrapper[4706]: I1127 07:18:57.949859 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" event={"ID":"19aee991-c9f9-4d97-9af1-3162462f7d1c","Type":"ContainerDied","Data":"f4af92f2b0e99f6e05936eeefe70f3496c19f7f4a87e3b123f7ca89258476c0a"} Nov 27 07:18:57 crc kubenswrapper[4706]: I1127 07:18:57.952195 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 07:18:59 crc kubenswrapper[4706]: I1127 07:18:59.963931 4706 generic.go:334] "Generic (PLEG): container finished" podID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerID="daf233511aa9de0ba70f30ebad8fa27bd0f03b50db3d1e4305a618a34fdf5a16" exitCode=0 Nov 27 07:18:59 crc kubenswrapper[4706]: I1127 07:18:59.964431 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" event={"ID":"19aee991-c9f9-4d97-9af1-3162462f7d1c","Type":"ContainerDied","Data":"daf233511aa9de0ba70f30ebad8fa27bd0f03b50db3d1e4305a618a34fdf5a16"} Nov 27 07:19:00 crc kubenswrapper[4706]: I1127 07:19:00.976096 4706 generic.go:334] "Generic (PLEG): container finished" podID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerID="a7742777b1e9f2abd15151a986d7c9475529a882f5ea0e404c2cbda699c000ca" exitCode=0 Nov 27 07:19:00 crc kubenswrapper[4706]: I1127 07:19:00.976210 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" event={"ID":"19aee991-c9f9-4d97-9af1-3162462f7d1c","Type":"ContainerDied","Data":"a7742777b1e9f2abd15151a986d7c9475529a882f5ea0e404c2cbda699c000ca"} Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.206420 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.329585 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw8ml\" (UniqueName: \"kubernetes.io/projected/19aee991-c9f9-4d97-9af1-3162462f7d1c-kube-api-access-lw8ml\") pod \"19aee991-c9f9-4d97-9af1-3162462f7d1c\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.329768 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-util\") pod \"19aee991-c9f9-4d97-9af1-3162462f7d1c\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.329811 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-bundle\") pod \"19aee991-c9f9-4d97-9af1-3162462f7d1c\" (UID: \"19aee991-c9f9-4d97-9af1-3162462f7d1c\") " Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.331461 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-bundle" (OuterVolumeSpecName: "bundle") pod "19aee991-c9f9-4d97-9af1-3162462f7d1c" (UID: "19aee991-c9f9-4d97-9af1-3162462f7d1c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.335650 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19aee991-c9f9-4d97-9af1-3162462f7d1c-kube-api-access-lw8ml" (OuterVolumeSpecName: "kube-api-access-lw8ml") pod "19aee991-c9f9-4d97-9af1-3162462f7d1c" (UID: "19aee991-c9f9-4d97-9af1-3162462f7d1c"). InnerVolumeSpecName "kube-api-access-lw8ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.355827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-util" (OuterVolumeSpecName: "util") pod "19aee991-c9f9-4d97-9af1-3162462f7d1c" (UID: "19aee991-c9f9-4d97-9af1-3162462f7d1c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.431212 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.431265 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19aee991-c9f9-4d97-9af1-3162462f7d1c-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.431277 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw8ml\" (UniqueName: \"kubernetes.io/projected/19aee991-c9f9-4d97-9af1-3162462f7d1c-kube-api-access-lw8ml\") on node \"crc\" DevicePath \"\"" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.992645 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" event={"ID":"19aee991-c9f9-4d97-9af1-3162462f7d1c","Type":"ContainerDied","Data":"7e915de1cdfff0ef18179e5760beed1c63e3cbf7c5de5111a9f3876794a1c9a5"} Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.992698 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e915de1cdfff0ef18179e5760beed1c63e3cbf7c5de5111a9f3876794a1c9a5" Nov 27 07:19:02 crc kubenswrapper[4706]: I1127 07:19:02.992709 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.641413 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5"] Nov 27 07:19:11 crc kubenswrapper[4706]: E1127 07:19:11.642124 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerName="extract" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.642142 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerName="extract" Nov 27 07:19:11 crc kubenswrapper[4706]: E1127 07:19:11.642165 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerName="pull" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.642173 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerName="pull" Nov 27 07:19:11 crc kubenswrapper[4706]: E1127 07:19:11.642186 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerName="util" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.642195 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerName="util" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.642329 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="19aee991-c9f9-4d97-9af1-3162462f7d1c" containerName="extract" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.642780 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.644465 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.644664 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.645962 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.646071 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-4knl6" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.646136 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.653532 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5"] Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.744158 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4941533e-f387-4b8b-96fd-933271e8f532-apiservice-cert\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.744210 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkvd9\" (UniqueName: \"kubernetes.io/projected/4941533e-f387-4b8b-96fd-933271e8f532-kube-api-access-wkvd9\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.744245 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4941533e-f387-4b8b-96fd-933271e8f532-webhook-cert\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.845708 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4941533e-f387-4b8b-96fd-933271e8f532-webhook-cert\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.845813 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4941533e-f387-4b8b-96fd-933271e8f532-apiservice-cert\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.845847 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkvd9\" (UniqueName: \"kubernetes.io/projected/4941533e-f387-4b8b-96fd-933271e8f532-kube-api-access-wkvd9\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.859043 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4941533e-f387-4b8b-96fd-933271e8f532-apiservice-cert\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.859505 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4941533e-f387-4b8b-96fd-933271e8f532-webhook-cert\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.871384 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkvd9\" (UniqueName: \"kubernetes.io/projected/4941533e-f387-4b8b-96fd-933271e8f532-kube-api-access-wkvd9\") pod \"metallb-operator-controller-manager-79b86cb64-xd9k5\" (UID: \"4941533e-f387-4b8b-96fd-933271e8f532\") " pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.901648 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76"] Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.902447 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.904471 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pg4tb" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.905064 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.905331 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.923545 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76"] Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.946565 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73781e8a-5a08-4e58-b974-2c9ef4242257-webhook-cert\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.946848 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d86pq\" (UniqueName: \"kubernetes.io/projected/73781e8a-5a08-4e58-b974-2c9ef4242257-kube-api-access-d86pq\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.946882 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73781e8a-5a08-4e58-b974-2c9ef4242257-apiservice-cert\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:11 crc kubenswrapper[4706]: I1127 07:19:11.957475 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.058602 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d86pq\" (UniqueName: \"kubernetes.io/projected/73781e8a-5a08-4e58-b974-2c9ef4242257-kube-api-access-d86pq\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.058677 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73781e8a-5a08-4e58-b974-2c9ef4242257-apiservice-cert\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.058803 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73781e8a-5a08-4e58-b974-2c9ef4242257-webhook-cert\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.069007 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73781e8a-5a08-4e58-b974-2c9ef4242257-apiservice-cert\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.071676 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d86pq\" (UniqueName: \"kubernetes.io/projected/73781e8a-5a08-4e58-b974-2c9ef4242257-kube-api-access-d86pq\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.077556 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73781e8a-5a08-4e58-b974-2c9ef4242257-webhook-cert\") pod \"metallb-operator-webhook-server-69f857b4f-nkg76\" (UID: \"73781e8a-5a08-4e58-b974-2c9ef4242257\") " pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.169034 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5"] Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.230859 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:12 crc kubenswrapper[4706]: I1127 07:19:12.448050 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76"] Nov 27 07:19:12 crc kubenswrapper[4706]: W1127 07:19:12.452424 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73781e8a_5a08_4e58_b974_2c9ef4242257.slice/crio-54af1f83075839ecdce20dd2a79a6a9e54b0f3d14be16845227d01b4265ff986 WatchSource:0}: Error finding container 54af1f83075839ecdce20dd2a79a6a9e54b0f3d14be16845227d01b4265ff986: Status 404 returned error can't find the container with id 54af1f83075839ecdce20dd2a79a6a9e54b0f3d14be16845227d01b4265ff986 Nov 27 07:19:13 crc kubenswrapper[4706]: I1127 07:19:13.045899 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" event={"ID":"73781e8a-5a08-4e58-b974-2c9ef4242257","Type":"ContainerStarted","Data":"54af1f83075839ecdce20dd2a79a6a9e54b0f3d14be16845227d01b4265ff986"} Nov 27 07:19:13 crc kubenswrapper[4706]: I1127 07:19:13.051255 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" event={"ID":"4941533e-f387-4b8b-96fd-933271e8f532","Type":"ContainerStarted","Data":"3ba2272e416759ec49760a04708190d89b1de25c30ee2e94acc2a2f5c2750447"} Nov 27 07:19:18 crc kubenswrapper[4706]: I1127 07:19:18.539961 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" event={"ID":"4941533e-f387-4b8b-96fd-933271e8f532","Type":"ContainerStarted","Data":"77c76d968a6cea9715bf955b1a77fe54616dc91970f712e898ee0e85384d6e9e"} Nov 27 07:19:18 crc kubenswrapper[4706]: I1127 07:19:18.540497 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:18 crc kubenswrapper[4706]: I1127 07:19:18.541967 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" event={"ID":"73781e8a-5a08-4e58-b974-2c9ef4242257","Type":"ContainerStarted","Data":"cea37af5f53aafab3aa4e6e1ffda15483c1805d0473fcde95a04eb2812bf88be"} Nov 27 07:19:18 crc kubenswrapper[4706]: I1127 07:19:18.542127 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:18 crc kubenswrapper[4706]: I1127 07:19:18.577800 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" podStartSLOduration=2.699157946 podStartE2EDuration="7.577786994s" podCreationTimestamp="2025-11-27 07:19:11 +0000 UTC" firstStartedPulling="2025-11-27 07:19:12.182776671 +0000 UTC m=+636.072367481" lastFinishedPulling="2025-11-27 07:19:17.061405719 +0000 UTC m=+640.950996529" observedRunningTime="2025-11-27 07:19:18.57392477 +0000 UTC m=+642.463515580" watchObservedRunningTime="2025-11-27 07:19:18.577786994 +0000 UTC m=+642.467377804" Nov 27 07:19:18 crc kubenswrapper[4706]: I1127 07:19:18.597073 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" podStartSLOduration=2.9872362040000002 podStartE2EDuration="7.597057631s" podCreationTimestamp="2025-11-27 07:19:11 +0000 UTC" firstStartedPulling="2025-11-27 07:19:12.456770009 +0000 UTC m=+636.346360819" lastFinishedPulling="2025-11-27 07:19:17.066591426 +0000 UTC m=+640.956182246" observedRunningTime="2025-11-27 07:19:18.593013861 +0000 UTC m=+642.482604671" watchObservedRunningTime="2025-11-27 07:19:18.597057631 +0000 UTC m=+642.486648441" Nov 27 07:19:32 crc kubenswrapper[4706]: I1127 07:19:32.236481 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-69f857b4f-nkg76" Nov 27 07:19:51 crc kubenswrapper[4706]: I1127 07:19:51.959792 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-79b86cb64-xd9k5" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.601164 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bkj8q"] Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.604720 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.608941 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22"] Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.610052 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.613117 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.613455 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.614105 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-h6j94" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.614343 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.627845 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22"] Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.647705 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-sockets\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.647759 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-startup\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.647789 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfhlp\" (UniqueName: \"kubernetes.io/projected/b893961c-2ac3-4a53-86c3-ffc1529ac013-kube-api-access-jfhlp\") pod \"frr-k8s-webhook-server-7fcb986d4-6lz22\" (UID: \"b893961c-2ac3-4a53-86c3-ffc1529ac013\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.648018 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.648153 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc9s2\" (UniqueName: \"kubernetes.io/projected/de4c78be-fea0-452b-bfb7-59b423f0cdf7-kube-api-access-sc9s2\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.648258 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-reloader\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.648332 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics-certs\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.648479 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-conf\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.648532 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b893961c-2ac3-4a53-86c3-ffc1529ac013-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6lz22\" (UID: \"b893961c-2ac3-4a53-86c3-ffc1529ac013\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.683976 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-2r7r8"] Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.685327 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.687272 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.687591 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kg26g" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.687728 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.688263 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.701508 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-58swk"] Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.702638 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.704412 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.722039 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-58swk"] Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.765918 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-metrics-certs\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.765967 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-cert\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.765988 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h26xl\" (UniqueName: \"kubernetes.io/projected/456beb89-e656-49ac-b6d2-ad92029ae549-kube-api-access-h26xl\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766023 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-conf\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766043 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b893961c-2ac3-4a53-86c3-ffc1529ac013-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6lz22\" (UID: \"b893961c-2ac3-4a53-86c3-ffc1529ac013\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766063 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-metrics-certs\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766446 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-sockets\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766543 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-conf\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766585 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/456beb89-e656-49ac-b6d2-ad92029ae549-metallb-excludel2\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766665 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-startup\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766732 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766788 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfhlp\" (UniqueName: \"kubernetes.io/projected/b893961c-2ac3-4a53-86c3-ffc1529ac013-kube-api-access-jfhlp\") pod \"frr-k8s-webhook-server-7fcb986d4-6lz22\" (UID: \"b893961c-2ac3-4a53-86c3-ffc1529ac013\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766808 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-sockets\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766891 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwx9m\" (UniqueName: \"kubernetes.io/projected/d987dc12-22dc-43cb-95dd-c40d07ca3786-kube-api-access-nwx9m\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766923 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766967 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc9s2\" (UniqueName: \"kubernetes.io/projected/de4c78be-fea0-452b-bfb7-59b423f0cdf7-kube-api-access-sc9s2\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.766996 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-reloader\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.767031 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics-certs\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: E1127 07:19:52.767184 4706 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 27 07:19:52 crc kubenswrapper[4706]: E1127 07:19:52.767264 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics-certs podName:de4c78be-fea0-452b-bfb7-59b423f0cdf7 nodeName:}" failed. No retries permitted until 2025-11-27 07:19:53.267242259 +0000 UTC m=+677.156833069 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics-certs") pod "frr-k8s-bkj8q" (UID: "de4c78be-fea0-452b-bfb7-59b423f0cdf7") : secret "frr-k8s-certs-secret" not found Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.767419 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/de4c78be-fea0-452b-bfb7-59b423f0cdf7-frr-startup\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.767660 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.767697 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/de4c78be-fea0-452b-bfb7-59b423f0cdf7-reloader\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.774875 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b893961c-2ac3-4a53-86c3-ffc1529ac013-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6lz22\" (UID: \"b893961c-2ac3-4a53-86c3-ffc1529ac013\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.799058 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfhlp\" (UniqueName: \"kubernetes.io/projected/b893961c-2ac3-4a53-86c3-ffc1529ac013-kube-api-access-jfhlp\") pod \"frr-k8s-webhook-server-7fcb986d4-6lz22\" (UID: \"b893961c-2ac3-4a53-86c3-ffc1529ac013\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.805932 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc9s2\" (UniqueName: \"kubernetes.io/projected/de4c78be-fea0-452b-bfb7-59b423f0cdf7-kube-api-access-sc9s2\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.868574 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-metrics-certs\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.868623 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-cert\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.868645 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h26xl\" (UniqueName: \"kubernetes.io/projected/456beb89-e656-49ac-b6d2-ad92029ae549-kube-api-access-h26xl\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.868730 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-metrics-certs\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: E1127 07:19:52.868734 4706 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 27 07:19:52 crc kubenswrapper[4706]: E1127 07:19:52.868802 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-metrics-certs podName:d987dc12-22dc-43cb-95dd-c40d07ca3786 nodeName:}" failed. No retries permitted until 2025-11-27 07:19:53.368785735 +0000 UTC m=+677.258376545 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-metrics-certs") pod "controller-f8648f98b-58swk" (UID: "d987dc12-22dc-43cb-95dd-c40d07ca3786") : secret "controller-certs-secret" not found Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.868829 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/456beb89-e656-49ac-b6d2-ad92029ae549-metallb-excludel2\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.868875 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.868924 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwx9m\" (UniqueName: \"kubernetes.io/projected/d987dc12-22dc-43cb-95dd-c40d07ca3786-kube-api-access-nwx9m\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: E1127 07:19:52.869196 4706 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 07:19:52 crc kubenswrapper[4706]: E1127 07:19:52.869285 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist podName:456beb89-e656-49ac-b6d2-ad92029ae549 nodeName:}" failed. No retries permitted until 2025-11-27 07:19:53.36926688 +0000 UTC m=+677.258857690 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist") pod "speaker-2r7r8" (UID: "456beb89-e656-49ac-b6d2-ad92029ae549") : secret "metallb-memberlist" not found Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.870117 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/456beb89-e656-49ac-b6d2-ad92029ae549-metallb-excludel2\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.871934 4706 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.871974 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-metrics-certs\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.885672 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-cert\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.894027 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h26xl\" (UniqueName: \"kubernetes.io/projected/456beb89-e656-49ac-b6d2-ad92029ae549-kube-api-access-h26xl\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.896045 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwx9m\" (UniqueName: \"kubernetes.io/projected/d987dc12-22dc-43cb-95dd-c40d07ca3786-kube-api-access-nwx9m\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:52 crc kubenswrapper[4706]: I1127 07:19:52.952582 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.127886 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22"] Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.273875 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics-certs\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.277378 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de4c78be-fea0-452b-bfb7-59b423f0cdf7-metrics-certs\") pod \"frr-k8s-bkj8q\" (UID: \"de4c78be-fea0-452b-bfb7-59b423f0cdf7\") " pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.375327 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-metrics-certs\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.375424 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:53 crc kubenswrapper[4706]: E1127 07:19:53.375584 4706 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 07:19:53 crc kubenswrapper[4706]: E1127 07:19:53.375651 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist podName:456beb89-e656-49ac-b6d2-ad92029ae549 nodeName:}" failed. No retries permitted until 2025-11-27 07:19:54.375633193 +0000 UTC m=+678.265224003 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist") pod "speaker-2r7r8" (UID: "456beb89-e656-49ac-b6d2-ad92029ae549") : secret "metallb-memberlist" not found Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.378368 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d987dc12-22dc-43cb-95dd-c40d07ca3786-metrics-certs\") pod \"controller-f8648f98b-58swk\" (UID: \"d987dc12-22dc-43cb-95dd-c40d07ca3786\") " pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.532142 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.672314 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.730556 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerStarted","Data":"1ace55a9e224c1250785828008b7dffc3d8528c02bc492880a49b099cd756dd4"} Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.732377 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" event={"ID":"b893961c-2ac3-4a53-86c3-ffc1529ac013","Type":"ContainerStarted","Data":"3f16d14d5afdf60ff679f1f6bfaed4232db507f303dc8c380362c0dcb44cd3a2"} Nov 27 07:19:53 crc kubenswrapper[4706]: I1127 07:19:53.931162 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-58swk"] Nov 27 07:19:53 crc kubenswrapper[4706]: W1127 07:19:53.937602 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd987dc12_22dc_43cb_95dd_c40d07ca3786.slice/crio-b1ee4636fbc9f6e3c5c76e003a0d8ecf248cfcf25c24c10e7c894e9f8f0dc93e WatchSource:0}: Error finding container b1ee4636fbc9f6e3c5c76e003a0d8ecf248cfcf25c24c10e7c894e9f8f0dc93e: Status 404 returned error can't find the container with id b1ee4636fbc9f6e3c5c76e003a0d8ecf248cfcf25c24c10e7c894e9f8f0dc93e Nov 27 07:19:54 crc kubenswrapper[4706]: I1127 07:19:54.388702 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:54 crc kubenswrapper[4706]: I1127 07:19:54.401468 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/456beb89-e656-49ac-b6d2-ad92029ae549-memberlist\") pod \"speaker-2r7r8\" (UID: \"456beb89-e656-49ac-b6d2-ad92029ae549\") " pod="metallb-system/speaker-2r7r8" Nov 27 07:19:54 crc kubenswrapper[4706]: I1127 07:19:54.515571 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-2r7r8" Nov 27 07:19:54 crc kubenswrapper[4706]: W1127 07:19:54.537991 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod456beb89_e656_49ac_b6d2_ad92029ae549.slice/crio-62ef8a28e6e018e8d39dba07264507860f2be2bbe8eeb81ac9fa46b2c8d836c4 WatchSource:0}: Error finding container 62ef8a28e6e018e8d39dba07264507860f2be2bbe8eeb81ac9fa46b2c8d836c4: Status 404 returned error can't find the container with id 62ef8a28e6e018e8d39dba07264507860f2be2bbe8eeb81ac9fa46b2c8d836c4 Nov 27 07:19:54 crc kubenswrapper[4706]: I1127 07:19:54.741897 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-58swk" event={"ID":"d987dc12-22dc-43cb-95dd-c40d07ca3786","Type":"ContainerStarted","Data":"d9859626276eeb9608cdcb0c94ed92e63a894ef229a5de8f30caba9bafa2b791"} Nov 27 07:19:54 crc kubenswrapper[4706]: I1127 07:19:54.741980 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-58swk" event={"ID":"d987dc12-22dc-43cb-95dd-c40d07ca3786","Type":"ContainerStarted","Data":"b1ee4636fbc9f6e3c5c76e003a0d8ecf248cfcf25c24c10e7c894e9f8f0dc93e"} Nov 27 07:19:54 crc kubenswrapper[4706]: I1127 07:19:54.742984 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2r7r8" event={"ID":"456beb89-e656-49ac-b6d2-ad92029ae549","Type":"ContainerStarted","Data":"62ef8a28e6e018e8d39dba07264507860f2be2bbe8eeb81ac9fa46b2c8d836c4"} Nov 27 07:19:55 crc kubenswrapper[4706]: I1127 07:19:55.753818 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2r7r8" event={"ID":"456beb89-e656-49ac-b6d2-ad92029ae549","Type":"ContainerStarted","Data":"b1f1996f3fb3b0f7ede05a3a5e0612bce755dcd0d5a71e9e9a8100b163079d2c"} Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.793279 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" event={"ID":"b893961c-2ac3-4a53-86c3-ffc1529ac013","Type":"ContainerStarted","Data":"7600710d4d00155b1e5de69cd528bbdd380874ed4a2dd7cffb6c57d550a558ca"} Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.793871 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.795549 4706 generic.go:334] "Generic (PLEG): container finished" podID="de4c78be-fea0-452b-bfb7-59b423f0cdf7" containerID="bca542fc6cec4515040e64f4a05aa7814d0f58d81a32730b7699926cc9c0aba6" exitCode=0 Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.795615 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerDied","Data":"bca542fc6cec4515040e64f4a05aa7814d0f58d81a32730b7699926cc9c0aba6"} Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.801300 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2r7r8" event={"ID":"456beb89-e656-49ac-b6d2-ad92029ae549","Type":"ContainerStarted","Data":"f607dd4d598b5e2ac0203a380a09b0045e766f5be73a0b579bcb31bc3c149220"} Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.802010 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-2r7r8" Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.805843 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-58swk" event={"ID":"d987dc12-22dc-43cb-95dd-c40d07ca3786","Type":"ContainerStarted","Data":"c64db67b830fe5e01dfd54762fa24ce0db4f993c75bf00c10d0ae7bfb8ca258b"} Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.806737 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.833134 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" podStartSLOduration=2.113933161 podStartE2EDuration="8.833112923s" podCreationTimestamp="2025-11-27 07:19:52 +0000 UTC" firstStartedPulling="2025-11-27 07:19:53.142936523 +0000 UTC m=+677.032527333" lastFinishedPulling="2025-11-27 07:19:59.862116285 +0000 UTC m=+683.751707095" observedRunningTime="2025-11-27 07:20:00.828788555 +0000 UTC m=+684.718379365" watchObservedRunningTime="2025-11-27 07:20:00.833112923 +0000 UTC m=+684.722703733" Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.860470 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-58swk" podStartSLOduration=3.073160083 podStartE2EDuration="8.86044603s" podCreationTimestamp="2025-11-27 07:19:52 +0000 UTC" firstStartedPulling="2025-11-27 07:19:54.075450058 +0000 UTC m=+677.965040868" lastFinishedPulling="2025-11-27 07:19:59.862736005 +0000 UTC m=+683.752326815" observedRunningTime="2025-11-27 07:20:00.857050511 +0000 UTC m=+684.746641331" watchObservedRunningTime="2025-11-27 07:20:00.86044603 +0000 UTC m=+684.750036840" Nov 27 07:20:00 crc kubenswrapper[4706]: I1127 07:20:00.889531 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-2r7r8" podStartSLOduration=3.795159127 podStartE2EDuration="8.889509931s" podCreationTimestamp="2025-11-27 07:19:52 +0000 UTC" firstStartedPulling="2025-11-27 07:19:54.768906888 +0000 UTC m=+678.658497698" lastFinishedPulling="2025-11-27 07:19:59.863257692 +0000 UTC m=+683.752848502" observedRunningTime="2025-11-27 07:20:00.887465326 +0000 UTC m=+684.777056166" watchObservedRunningTime="2025-11-27 07:20:00.889509931 +0000 UTC m=+684.779100751" Nov 27 07:20:01 crc kubenswrapper[4706]: I1127 07:20:01.813143 4706 generic.go:334] "Generic (PLEG): container finished" podID="de4c78be-fea0-452b-bfb7-59b423f0cdf7" containerID="b6adca1879a9d8d382b3e87be4bee6dd65f5e4c17a9086f0889d11f9dd31093b" exitCode=0 Nov 27 07:20:01 crc kubenswrapper[4706]: I1127 07:20:01.813232 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerDied","Data":"b6adca1879a9d8d382b3e87be4bee6dd65f5e4c17a9086f0889d11f9dd31093b"} Nov 27 07:20:02 crc kubenswrapper[4706]: I1127 07:20:02.822424 4706 generic.go:334] "Generic (PLEG): container finished" podID="de4c78be-fea0-452b-bfb7-59b423f0cdf7" containerID="f1501bc585a6f9d3e7bc2bbd44b589021c4b5ec26688c953c00021befccb362c" exitCode=0 Nov 27 07:20:02 crc kubenswrapper[4706]: I1127 07:20:02.822516 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerDied","Data":"f1501bc585a6f9d3e7bc2bbd44b589021c4b5ec26688c953c00021befccb362c"} Nov 27 07:20:03 crc kubenswrapper[4706]: I1127 07:20:03.836800 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerStarted","Data":"659803b566764d27c9553976aeed04b61ae63c44191de6962e29a09c42496223"} Nov 27 07:20:03 crc kubenswrapper[4706]: I1127 07:20:03.837092 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerStarted","Data":"5a754a064f5d5bb3b9fdd65c5b6ae121cf69698873f60fa7ec2118ed650e2ea9"} Nov 27 07:20:03 crc kubenswrapper[4706]: I1127 07:20:03.837102 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerStarted","Data":"99de3d0605cc7a951075128dccd4413bf036a49763df5b511ef46bba9b6e69e4"} Nov 27 07:20:03 crc kubenswrapper[4706]: I1127 07:20:03.837110 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerStarted","Data":"63d14a96087b34e1137ac68859ab25e71ba8ef8229f55e7f3403b09dc4823d5e"} Nov 27 07:20:04 crc kubenswrapper[4706]: I1127 07:20:04.518810 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-2r7r8" Nov 27 07:20:04 crc kubenswrapper[4706]: I1127 07:20:04.855335 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerStarted","Data":"5f2bd8fa9010b9e6d90c7c357bb67ccf781fee34e8fb1514139ec69978f00369"} Nov 27 07:20:04 crc kubenswrapper[4706]: I1127 07:20:04.855727 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bkj8q" event={"ID":"de4c78be-fea0-452b-bfb7-59b423f0cdf7","Type":"ContainerStarted","Data":"98ba6aa3e3281bad277ecd088e5025a9888c7498ae4734523ccc9a6699972297"} Nov 27 07:20:04 crc kubenswrapper[4706]: I1127 07:20:04.855752 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:20:04 crc kubenswrapper[4706]: I1127 07:20:04.881587 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bkj8q" podStartSLOduration=6.6195077 podStartE2EDuration="12.881570068s" podCreationTimestamp="2025-11-27 07:19:52 +0000 UTC" firstStartedPulling="2025-11-27 07:19:53.635737201 +0000 UTC m=+677.525328011" lastFinishedPulling="2025-11-27 07:19:59.897799569 +0000 UTC m=+683.787390379" observedRunningTime="2025-11-27 07:20:04.878784078 +0000 UTC m=+688.768374888" watchObservedRunningTime="2025-11-27 07:20:04.881570068 +0000 UTC m=+688.771160878" Nov 27 07:20:08 crc kubenswrapper[4706]: I1127 07:20:08.532626 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:20:08 crc kubenswrapper[4706]: I1127 07:20:08.567912 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.357689 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-rjlrj"] Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.359000 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-rjlrj" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.360846 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.363089 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.363138 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-fq6j2" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.367618 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-rjlrj"] Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.492654 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dls2z\" (UniqueName: \"kubernetes.io/projected/40db01d7-6fea-44a1-8349-d64aec9194ac-kube-api-access-dls2z\") pod \"mariadb-operator-index-rjlrj\" (UID: \"40db01d7-6fea-44a1-8349-d64aec9194ac\") " pod="openstack-operators/mariadb-operator-index-rjlrj" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.594467 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dls2z\" (UniqueName: \"kubernetes.io/projected/40db01d7-6fea-44a1-8349-d64aec9194ac-kube-api-access-dls2z\") pod \"mariadb-operator-index-rjlrj\" (UID: \"40db01d7-6fea-44a1-8349-d64aec9194ac\") " pod="openstack-operators/mariadb-operator-index-rjlrj" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.615465 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dls2z\" (UniqueName: \"kubernetes.io/projected/40db01d7-6fea-44a1-8349-d64aec9194ac-kube-api-access-dls2z\") pod \"mariadb-operator-index-rjlrj\" (UID: \"40db01d7-6fea-44a1-8349-d64aec9194ac\") " pod="openstack-operators/mariadb-operator-index-rjlrj" Nov 27 07:20:10 crc kubenswrapper[4706]: I1127 07:20:10.675942 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-rjlrj" Nov 27 07:20:11 crc kubenswrapper[4706]: I1127 07:20:11.052074 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-rjlrj"] Nov 27 07:20:11 crc kubenswrapper[4706]: W1127 07:20:11.066570 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40db01d7_6fea_44a1_8349_d64aec9194ac.slice/crio-17543605ec9a575415f8a320b77fdc2864ee00d4873ed550691b0a73fa2b1695 WatchSource:0}: Error finding container 17543605ec9a575415f8a320b77fdc2864ee00d4873ed550691b0a73fa2b1695: Status 404 returned error can't find the container with id 17543605ec9a575415f8a320b77fdc2864ee00d4873ed550691b0a73fa2b1695 Nov 27 07:20:11 crc kubenswrapper[4706]: I1127 07:20:11.899663 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-rjlrj" event={"ID":"40db01d7-6fea-44a1-8349-d64aec9194ac","Type":"ContainerStarted","Data":"17543605ec9a575415f8a320b77fdc2864ee00d4873ed550691b0a73fa2b1695"} Nov 27 07:20:12 crc kubenswrapper[4706]: I1127 07:20:12.906540 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-rjlrj" event={"ID":"40db01d7-6fea-44a1-8349-d64aec9194ac","Type":"ContainerStarted","Data":"5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040"} Nov 27 07:20:12 crc kubenswrapper[4706]: I1127 07:20:12.924400 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-rjlrj" podStartSLOduration=2.073149683 podStartE2EDuration="2.924384162s" podCreationTimestamp="2025-11-27 07:20:10 +0000 UTC" firstStartedPulling="2025-11-27 07:20:11.068556248 +0000 UTC m=+694.958147058" lastFinishedPulling="2025-11-27 07:20:11.919790717 +0000 UTC m=+695.809381537" observedRunningTime="2025-11-27 07:20:12.920960803 +0000 UTC m=+696.810551643" watchObservedRunningTime="2025-11-27 07:20:12.924384162 +0000 UTC m=+696.813974972" Nov 27 07:20:12 crc kubenswrapper[4706]: I1127 07:20:12.958936 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6lz22" Nov 27 07:20:13 crc kubenswrapper[4706]: I1127 07:20:13.537923 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bkj8q" Nov 27 07:20:13 crc kubenswrapper[4706]: I1127 07:20:13.677865 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-58swk" Nov 27 07:20:13 crc kubenswrapper[4706]: I1127 07:20:13.725719 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-rjlrj"] Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.329854 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-5s7bl"] Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.330814 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.340663 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-5s7bl"] Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.452778 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kmng\" (UniqueName: \"kubernetes.io/projected/a5eff20a-aef4-43d4-b5a6-779aa196bdb3-kube-api-access-8kmng\") pod \"mariadb-operator-index-5s7bl\" (UID: \"a5eff20a-aef4-43d4-b5a6-779aa196bdb3\") " pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.553766 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kmng\" (UniqueName: \"kubernetes.io/projected/a5eff20a-aef4-43d4-b5a6-779aa196bdb3-kube-api-access-8kmng\") pod \"mariadb-operator-index-5s7bl\" (UID: \"a5eff20a-aef4-43d4-b5a6-779aa196bdb3\") " pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.570908 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kmng\" (UniqueName: \"kubernetes.io/projected/a5eff20a-aef4-43d4-b5a6-779aa196bdb3-kube-api-access-8kmng\") pod \"mariadb-operator-index-5s7bl\" (UID: \"a5eff20a-aef4-43d4-b5a6-779aa196bdb3\") " pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.657035 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.916029 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-rjlrj" podUID="40db01d7-6fea-44a1-8349-d64aec9194ac" containerName="registry-server" containerID="cri-o://5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040" gracePeriod=2 Nov 27 07:20:14 crc kubenswrapper[4706]: I1127 07:20:14.930079 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-5s7bl"] Nov 27 07:20:14 crc kubenswrapper[4706]: W1127 07:20:14.944070 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5eff20a_aef4_43d4_b5a6_779aa196bdb3.slice/crio-f49f7303c56ca132c8f770e36263041388f5d231f489f2d7354601f4c064a76a WatchSource:0}: Error finding container f49f7303c56ca132c8f770e36263041388f5d231f489f2d7354601f4c064a76a: Status 404 returned error can't find the container with id f49f7303c56ca132c8f770e36263041388f5d231f489f2d7354601f4c064a76a Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.177853 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.178262 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.223459 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-rjlrj" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.374547 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dls2z\" (UniqueName: \"kubernetes.io/projected/40db01d7-6fea-44a1-8349-d64aec9194ac-kube-api-access-dls2z\") pod \"40db01d7-6fea-44a1-8349-d64aec9194ac\" (UID: \"40db01d7-6fea-44a1-8349-d64aec9194ac\") " Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.379061 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40db01d7-6fea-44a1-8349-d64aec9194ac-kube-api-access-dls2z" (OuterVolumeSpecName: "kube-api-access-dls2z") pod "40db01d7-6fea-44a1-8349-d64aec9194ac" (UID: "40db01d7-6fea-44a1-8349-d64aec9194ac"). InnerVolumeSpecName "kube-api-access-dls2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.476319 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dls2z\" (UniqueName: \"kubernetes.io/projected/40db01d7-6fea-44a1-8349-d64aec9194ac-kube-api-access-dls2z\") on node \"crc\" DevicePath \"\"" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.923064 4706 generic.go:334] "Generic (PLEG): container finished" podID="40db01d7-6fea-44a1-8349-d64aec9194ac" containerID="5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040" exitCode=0 Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.923121 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-rjlrj" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.923134 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-rjlrj" event={"ID":"40db01d7-6fea-44a1-8349-d64aec9194ac","Type":"ContainerDied","Data":"5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040"} Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.923165 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-rjlrj" event={"ID":"40db01d7-6fea-44a1-8349-d64aec9194ac","Type":"ContainerDied","Data":"17543605ec9a575415f8a320b77fdc2864ee00d4873ed550691b0a73fa2b1695"} Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.923182 4706 scope.go:117] "RemoveContainer" containerID="5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.924658 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5s7bl" event={"ID":"a5eff20a-aef4-43d4-b5a6-779aa196bdb3","Type":"ContainerStarted","Data":"feaac72fbe9a0b58641ccab10d5f231dd867b0b07582a01647eb3e0e7a608832"} Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.924697 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5s7bl" event={"ID":"a5eff20a-aef4-43d4-b5a6-779aa196bdb3","Type":"ContainerStarted","Data":"f49f7303c56ca132c8f770e36263041388f5d231f489f2d7354601f4c064a76a"} Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.939501 4706 scope.go:117] "RemoveContainer" containerID="5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040" Nov 27 07:20:15 crc kubenswrapper[4706]: E1127 07:20:15.940076 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040\": container with ID starting with 5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040 not found: ID does not exist" containerID="5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.940136 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040"} err="failed to get container status \"5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040\": rpc error: code = NotFound desc = could not find container \"5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040\": container with ID starting with 5f95e47826ae13283195de87e7ac8715c17f879e317de1a69b6ba610c7b58040 not found: ID does not exist" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.967356 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-5s7bl" podStartSLOduration=1.455958047 podStartE2EDuration="1.967327571s" podCreationTimestamp="2025-11-27 07:20:14 +0000 UTC" firstStartedPulling="2025-11-27 07:20:14.948956495 +0000 UTC m=+698.838547315" lastFinishedPulling="2025-11-27 07:20:15.460326029 +0000 UTC m=+699.349916839" observedRunningTime="2025-11-27 07:20:15.950532053 +0000 UTC m=+699.840122903" watchObservedRunningTime="2025-11-27 07:20:15.967327571 +0000 UTC m=+699.856918421" Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.971186 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-rjlrj"] Nov 27 07:20:15 crc kubenswrapper[4706]: I1127 07:20:15.977976 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-rjlrj"] Nov 27 07:20:16 crc kubenswrapper[4706]: I1127 07:20:16.788108 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40db01d7-6fea-44a1-8349-d64aec9194ac" path="/var/lib/kubelet/pods/40db01d7-6fea-44a1-8349-d64aec9194ac/volumes" Nov 27 07:20:24 crc kubenswrapper[4706]: I1127 07:20:24.657596 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:24 crc kubenswrapper[4706]: I1127 07:20:24.657901 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:24 crc kubenswrapper[4706]: I1127 07:20:24.701371 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:25 crc kubenswrapper[4706]: I1127 07:20:25.097901 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.390573 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg"] Nov 27 07:20:26 crc kubenswrapper[4706]: E1127 07:20:26.391599 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40db01d7-6fea-44a1-8349-d64aec9194ac" containerName="registry-server" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.391634 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="40db01d7-6fea-44a1-8349-d64aec9194ac" containerName="registry-server" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.391935 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="40db01d7-6fea-44a1-8349-d64aec9194ac" containerName="registry-server" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.393657 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.403529 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dk9r5" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.406092 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg"] Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.529201 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.529410 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn5jq\" (UniqueName: \"kubernetes.io/projected/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-kube-api-access-hn5jq\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.529897 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.631003 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn5jq\" (UniqueName: \"kubernetes.io/projected/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-kube-api-access-hn5jq\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.631081 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.631176 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.631690 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.632155 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.659337 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn5jq\" (UniqueName: \"kubernetes.io/projected/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-kube-api-access-hn5jq\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:26 crc kubenswrapper[4706]: I1127 07:20:26.721558 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:27 crc kubenswrapper[4706]: I1127 07:20:27.203043 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg"] Nov 27 07:20:27 crc kubenswrapper[4706]: W1127 07:20:27.213516 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b5ddba9_e78b_486e_9b93_a2ab98e0a09f.slice/crio-e7b4049163ea236e8731be65e10947457064c93b680a4daad1c21931c8f349e6 WatchSource:0}: Error finding container e7b4049163ea236e8731be65e10947457064c93b680a4daad1c21931c8f349e6: Status 404 returned error can't find the container with id e7b4049163ea236e8731be65e10947457064c93b680a4daad1c21931c8f349e6 Nov 27 07:20:28 crc kubenswrapper[4706]: I1127 07:20:28.079479 4706 generic.go:334] "Generic (PLEG): container finished" podID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerID="a9b8ab89a86be81768bb0ca0e264236bebf22fb19be64499857af94a9d8d3032" exitCode=0 Nov 27 07:20:28 crc kubenswrapper[4706]: I1127 07:20:28.079550 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" event={"ID":"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f","Type":"ContainerDied","Data":"a9b8ab89a86be81768bb0ca0e264236bebf22fb19be64499857af94a9d8d3032"} Nov 27 07:20:28 crc kubenswrapper[4706]: I1127 07:20:28.079600 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" event={"ID":"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f","Type":"ContainerStarted","Data":"e7b4049163ea236e8731be65e10947457064c93b680a4daad1c21931c8f349e6"} Nov 27 07:20:29 crc kubenswrapper[4706]: I1127 07:20:29.086801 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" event={"ID":"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f","Type":"ContainerStarted","Data":"3c7cb477052ba49d8ac443222892feee698a2a40cb1a09f9c2f5731a541c1c07"} Nov 27 07:20:30 crc kubenswrapper[4706]: I1127 07:20:30.094920 4706 generic.go:334] "Generic (PLEG): container finished" podID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerID="3c7cb477052ba49d8ac443222892feee698a2a40cb1a09f9c2f5731a541c1c07" exitCode=0 Nov 27 07:20:30 crc kubenswrapper[4706]: I1127 07:20:30.094960 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" event={"ID":"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f","Type":"ContainerDied","Data":"3c7cb477052ba49d8ac443222892feee698a2a40cb1a09f9c2f5731a541c1c07"} Nov 27 07:20:31 crc kubenswrapper[4706]: I1127 07:20:31.101455 4706 generic.go:334] "Generic (PLEG): container finished" podID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerID="dd4684d059a4f4303d5e038b480569409a65ff70f6401f9cafdad71df6843d5a" exitCode=0 Nov 27 07:20:31 crc kubenswrapper[4706]: I1127 07:20:31.101488 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" event={"ID":"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f","Type":"ContainerDied","Data":"dd4684d059a4f4303d5e038b480569409a65ff70f6401f9cafdad71df6843d5a"} Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.380663 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.518330 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn5jq\" (UniqueName: \"kubernetes.io/projected/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-kube-api-access-hn5jq\") pod \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.518471 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-bundle\") pod \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.519365 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-util\") pod \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\" (UID: \"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f\") " Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.520804 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-bundle" (OuterVolumeSpecName: "bundle") pod "8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" (UID: "8b5ddba9-e78b-486e-9b93-a2ab98e0a09f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.525168 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-kube-api-access-hn5jq" (OuterVolumeSpecName: "kube-api-access-hn5jq") pod "8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" (UID: "8b5ddba9-e78b-486e-9b93-a2ab98e0a09f"). InnerVolumeSpecName "kube-api-access-hn5jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.539110 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-util" (OuterVolumeSpecName: "util") pod "8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" (UID: "8b5ddba9-e78b-486e-9b93-a2ab98e0a09f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.620479 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.620530 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn5jq\" (UniqueName: \"kubernetes.io/projected/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-kube-api-access-hn5jq\") on node \"crc\" DevicePath \"\"" Nov 27 07:20:32 crc kubenswrapper[4706]: I1127 07:20:32.620548 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:20:33 crc kubenswrapper[4706]: I1127 07:20:33.119547 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" event={"ID":"8b5ddba9-e78b-486e-9b93-a2ab98e0a09f","Type":"ContainerDied","Data":"e7b4049163ea236e8731be65e10947457064c93b680a4daad1c21931c8f349e6"} Nov 27 07:20:33 crc kubenswrapper[4706]: I1127 07:20:33.119582 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7b4049163ea236e8731be65e10947457064c93b680a4daad1c21931c8f349e6" Nov 27 07:20:33 crc kubenswrapper[4706]: I1127 07:20:33.119635 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.722418 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg"] Nov 27 07:20:39 crc kubenswrapper[4706]: E1127 07:20:39.723137 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerName="pull" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.723151 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerName="pull" Nov 27 07:20:39 crc kubenswrapper[4706]: E1127 07:20:39.723162 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerName="extract" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.723167 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerName="extract" Nov 27 07:20:39 crc kubenswrapper[4706]: E1127 07:20:39.723177 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerName="util" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.723183 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerName="util" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.723299 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" containerName="extract" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.723644 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.725969 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-dd6cn" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.726733 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.728710 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.738996 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg"] Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.819008 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-webhook-cert\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.819128 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7rjg\" (UniqueName: \"kubernetes.io/projected/4b1548e8-8230-4cd3-9579-ecc704681612-kube-api-access-c7rjg\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.819195 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-apiservice-cert\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.920057 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-apiservice-cert\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.920136 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-webhook-cert\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.920210 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7rjg\" (UniqueName: \"kubernetes.io/projected/4b1548e8-8230-4cd3-9579-ecc704681612-kube-api-access-c7rjg\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.929382 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-apiservice-cert\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.929480 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-webhook-cert\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:39 crc kubenswrapper[4706]: I1127 07:20:39.937846 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7rjg\" (UniqueName: \"kubernetes.io/projected/4b1548e8-8230-4cd3-9579-ecc704681612-kube-api-access-c7rjg\") pod \"mariadb-operator-controller-manager-759795887-m7ctg\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:40 crc kubenswrapper[4706]: I1127 07:20:40.041967 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:40 crc kubenswrapper[4706]: I1127 07:20:40.475450 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg"] Nov 27 07:20:41 crc kubenswrapper[4706]: I1127 07:20:41.170522 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" event={"ID":"4b1548e8-8230-4cd3-9579-ecc704681612","Type":"ContainerStarted","Data":"82eb6db289851aac5237d2a75165d37947c3f697215405b469b14f9eba88e75d"} Nov 27 07:20:44 crc kubenswrapper[4706]: I1127 07:20:44.188182 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" event={"ID":"4b1548e8-8230-4cd3-9579-ecc704681612","Type":"ContainerStarted","Data":"2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5"} Nov 27 07:20:44 crc kubenswrapper[4706]: I1127 07:20:44.188559 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:44 crc kubenswrapper[4706]: I1127 07:20:44.215384 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" podStartSLOduration=1.8323527849999999 podStartE2EDuration="5.215359145s" podCreationTimestamp="2025-11-27 07:20:39 +0000 UTC" firstStartedPulling="2025-11-27 07:20:40.484700719 +0000 UTC m=+724.374291539" lastFinishedPulling="2025-11-27 07:20:43.867707089 +0000 UTC m=+727.757297899" observedRunningTime="2025-11-27 07:20:44.210440267 +0000 UTC m=+728.100031107" watchObservedRunningTime="2025-11-27 07:20:44.215359145 +0000 UTC m=+728.104949975" Nov 27 07:20:45 crc kubenswrapper[4706]: I1127 07:20:45.178385 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:20:45 crc kubenswrapper[4706]: I1127 07:20:45.178726 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:20:50 crc kubenswrapper[4706]: I1127 07:20:50.050375 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:20:53 crc kubenswrapper[4706]: I1127 07:20:53.862318 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-86rtd"] Nov 27 07:20:53 crc kubenswrapper[4706]: I1127 07:20:53.863428 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-86rtd" Nov 27 07:20:53 crc kubenswrapper[4706]: I1127 07:20:53.865055 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-r5nbg" Nov 27 07:20:53 crc kubenswrapper[4706]: I1127 07:20:53.871813 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-86rtd"] Nov 27 07:20:53 crc kubenswrapper[4706]: I1127 07:20:53.988646 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj8kw\" (UniqueName: \"kubernetes.io/projected/dbcacf92-fdf4-4373-9c72-fa42ac6da370-kube-api-access-qj8kw\") pod \"infra-operator-index-86rtd\" (UID: \"dbcacf92-fdf4-4373-9c72-fa42ac6da370\") " pod="openstack-operators/infra-operator-index-86rtd" Nov 27 07:20:54 crc kubenswrapper[4706]: I1127 07:20:54.090061 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj8kw\" (UniqueName: \"kubernetes.io/projected/dbcacf92-fdf4-4373-9c72-fa42ac6da370-kube-api-access-qj8kw\") pod \"infra-operator-index-86rtd\" (UID: \"dbcacf92-fdf4-4373-9c72-fa42ac6da370\") " pod="openstack-operators/infra-operator-index-86rtd" Nov 27 07:20:54 crc kubenswrapper[4706]: I1127 07:20:54.108416 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj8kw\" (UniqueName: \"kubernetes.io/projected/dbcacf92-fdf4-4373-9c72-fa42ac6da370-kube-api-access-qj8kw\") pod \"infra-operator-index-86rtd\" (UID: \"dbcacf92-fdf4-4373-9c72-fa42ac6da370\") " pod="openstack-operators/infra-operator-index-86rtd" Nov 27 07:20:54 crc kubenswrapper[4706]: I1127 07:20:54.187568 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-86rtd" Nov 27 07:20:54 crc kubenswrapper[4706]: I1127 07:20:54.413738 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-86rtd"] Nov 27 07:20:55 crc kubenswrapper[4706]: I1127 07:20:55.257260 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-86rtd" event={"ID":"dbcacf92-fdf4-4373-9c72-fa42ac6da370","Type":"ContainerStarted","Data":"531cc1ff954f65c351e0afebd17df7c8e33a7493c32a670857ddb795b445af5d"} Nov 27 07:20:56 crc kubenswrapper[4706]: I1127 07:20:56.262518 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-86rtd" event={"ID":"dbcacf92-fdf4-4373-9c72-fa42ac6da370","Type":"ContainerStarted","Data":"b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466"} Nov 27 07:20:56 crc kubenswrapper[4706]: I1127 07:20:56.279460 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-86rtd" podStartSLOduration=2.383475827 podStartE2EDuration="3.279440746s" podCreationTimestamp="2025-11-27 07:20:53 +0000 UTC" firstStartedPulling="2025-11-27 07:20:54.425726241 +0000 UTC m=+738.315317061" lastFinishedPulling="2025-11-27 07:20:55.32169117 +0000 UTC m=+739.211281980" observedRunningTime="2025-11-27 07:20:56.272480663 +0000 UTC m=+740.162071483" watchObservedRunningTime="2025-11-27 07:20:56.279440746 +0000 UTC m=+740.169031556" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.261455 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-86rtd"] Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.274178 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-86rtd" podUID="dbcacf92-fdf4-4373-9c72-fa42ac6da370" containerName="registry-server" containerID="cri-o://b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466" gracePeriod=2 Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.650295 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-86rtd" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.749746 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj8kw\" (UniqueName: \"kubernetes.io/projected/dbcacf92-fdf4-4373-9c72-fa42ac6da370-kube-api-access-qj8kw\") pod \"dbcacf92-fdf4-4373-9c72-fa42ac6da370\" (UID: \"dbcacf92-fdf4-4373-9c72-fa42ac6da370\") " Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.760502 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbcacf92-fdf4-4373-9c72-fa42ac6da370-kube-api-access-qj8kw" (OuterVolumeSpecName: "kube-api-access-qj8kw") pod "dbcacf92-fdf4-4373-9c72-fa42ac6da370" (UID: "dbcacf92-fdf4-4373-9c72-fa42ac6da370"). InnerVolumeSpecName "kube-api-access-qj8kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.851047 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj8kw\" (UniqueName: \"kubernetes.io/projected/dbcacf92-fdf4-4373-9c72-fa42ac6da370-kube-api-access-qj8kw\") on node \"crc\" DevicePath \"\"" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.866251 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-ccm9g"] Nov 27 07:20:58 crc kubenswrapper[4706]: E1127 07:20:58.866533 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbcacf92-fdf4-4373-9c72-fa42ac6da370" containerName="registry-server" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.866556 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbcacf92-fdf4-4373-9c72-fa42ac6da370" containerName="registry-server" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.866670 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbcacf92-fdf4-4373-9c72-fa42ac6da370" containerName="registry-server" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.867238 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.880435 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-ccm9g"] Nov 27 07:20:58 crc kubenswrapper[4706]: I1127 07:20:58.952087 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2tbf\" (UniqueName: \"kubernetes.io/projected/5b31be76-beca-407c-a1a9-db1e8e3f100e-kube-api-access-s2tbf\") pod \"infra-operator-index-ccm9g\" (UID: \"5b31be76-beca-407c-a1a9-db1e8e3f100e\") " pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.054047 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2tbf\" (UniqueName: \"kubernetes.io/projected/5b31be76-beca-407c-a1a9-db1e8e3f100e-kube-api-access-s2tbf\") pod \"infra-operator-index-ccm9g\" (UID: \"5b31be76-beca-407c-a1a9-db1e8e3f100e\") " pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.068663 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2tbf\" (UniqueName: \"kubernetes.io/projected/5b31be76-beca-407c-a1a9-db1e8e3f100e-kube-api-access-s2tbf\") pod \"infra-operator-index-ccm9g\" (UID: \"5b31be76-beca-407c-a1a9-db1e8e3f100e\") " pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.184636 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.289099 4706 generic.go:334] "Generic (PLEG): container finished" podID="dbcacf92-fdf4-4373-9c72-fa42ac6da370" containerID="b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466" exitCode=0 Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.289194 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-86rtd" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.289190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-86rtd" event={"ID":"dbcacf92-fdf4-4373-9c72-fa42ac6da370","Type":"ContainerDied","Data":"b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466"} Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.289941 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-86rtd" event={"ID":"dbcacf92-fdf4-4373-9c72-fa42ac6da370","Type":"ContainerDied","Data":"531cc1ff954f65c351e0afebd17df7c8e33a7493c32a670857ddb795b445af5d"} Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.289962 4706 scope.go:117] "RemoveContainer" containerID="b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.313767 4706 scope.go:117] "RemoveContainer" containerID="b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466" Nov 27 07:20:59 crc kubenswrapper[4706]: E1127 07:20:59.314446 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466\": container with ID starting with b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466 not found: ID does not exist" containerID="b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.314493 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466"} err="failed to get container status \"b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466\": rpc error: code = NotFound desc = could not find container \"b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466\": container with ID starting with b4d043d0f9f9c32293cdcd249e6967400d4f0ee7f4e3ad9427d1d921c76cd466 not found: ID does not exist" Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.319027 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-86rtd"] Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.322609 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-86rtd"] Nov 27 07:20:59 crc kubenswrapper[4706]: I1127 07:20:59.386261 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-ccm9g"] Nov 27 07:20:59 crc kubenswrapper[4706]: W1127 07:20:59.390144 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b31be76_beca_407c_a1a9_db1e8e3f100e.slice/crio-c400281098673ce784a72f82e0b72e27861f8fd325b8715ecaaf1e0435e9aa91 WatchSource:0}: Error finding container c400281098673ce784a72f82e0b72e27861f8fd325b8715ecaaf1e0435e9aa91: Status 404 returned error can't find the container with id c400281098673ce784a72f82e0b72e27861f8fd325b8715ecaaf1e0435e9aa91 Nov 27 07:21:00 crc kubenswrapper[4706]: I1127 07:21:00.295935 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ccm9g" event={"ID":"5b31be76-beca-407c-a1a9-db1e8e3f100e","Type":"ContainerStarted","Data":"e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c"} Nov 27 07:21:00 crc kubenswrapper[4706]: I1127 07:21:00.296178 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ccm9g" event={"ID":"5b31be76-beca-407c-a1a9-db1e8e3f100e","Type":"ContainerStarted","Data":"c400281098673ce784a72f82e0b72e27861f8fd325b8715ecaaf1e0435e9aa91"} Nov 27 07:21:00 crc kubenswrapper[4706]: I1127 07:21:00.311125 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-ccm9g" podStartSLOduration=1.878919292 podStartE2EDuration="2.311106168s" podCreationTimestamp="2025-11-27 07:20:58 +0000 UTC" firstStartedPulling="2025-11-27 07:20:59.394212656 +0000 UTC m=+743.283803466" lastFinishedPulling="2025-11-27 07:20:59.826399522 +0000 UTC m=+743.715990342" observedRunningTime="2025-11-27 07:21:00.310075285 +0000 UTC m=+744.199666105" watchObservedRunningTime="2025-11-27 07:21:00.311106168 +0000 UTC m=+744.200696978" Nov 27 07:21:00 crc kubenswrapper[4706]: I1127 07:21:00.782557 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbcacf92-fdf4-4373-9c72-fa42ac6da370" path="/var/lib/kubelet/pods/dbcacf92-fdf4-4373-9c72-fa42ac6da370/volumes" Nov 27 07:21:09 crc kubenswrapper[4706]: I1127 07:21:09.185760 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:21:09 crc kubenswrapper[4706]: I1127 07:21:09.186293 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:21:09 crc kubenswrapper[4706]: I1127 07:21:09.213696 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:21:09 crc kubenswrapper[4706]: I1127 07:21:09.404615 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:21:10 crc kubenswrapper[4706]: I1127 07:21:10.895377 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr"] Nov 27 07:21:10 crc kubenswrapper[4706]: I1127 07:21:10.897732 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:10 crc kubenswrapper[4706]: I1127 07:21:10.900390 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dk9r5" Nov 27 07:21:10 crc kubenswrapper[4706]: I1127 07:21:10.925390 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr"] Nov 27 07:21:10 crc kubenswrapper[4706]: I1127 07:21:10.927049 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:10 crc kubenswrapper[4706]: I1127 07:21:10.927084 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:10 crc kubenswrapper[4706]: I1127 07:21:10.927146 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q47dq\" (UniqueName: \"kubernetes.io/projected/94169577-2199-446a-9d53-8473269fc045-kube-api-access-q47dq\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.028322 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q47dq\" (UniqueName: \"kubernetes.io/projected/94169577-2199-446a-9d53-8473269fc045-kube-api-access-q47dq\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.028799 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.028835 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.029426 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.029547 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.054241 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q47dq\" (UniqueName: \"kubernetes.io/projected/94169577-2199-446a-9d53-8473269fc045-kube-api-access-q47dq\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.212060 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.328639 4706 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 07:21:11 crc kubenswrapper[4706]: I1127 07:21:11.462874 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr"] Nov 27 07:21:11 crc kubenswrapper[4706]: W1127 07:21:11.466933 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94169577_2199_446a_9d53_8473269fc045.slice/crio-46eb2e1b6147319d52b871094fac9a43a335a7e20d5badd66498905f300e0413 WatchSource:0}: Error finding container 46eb2e1b6147319d52b871094fac9a43a335a7e20d5badd66498905f300e0413: Status 404 returned error can't find the container with id 46eb2e1b6147319d52b871094fac9a43a335a7e20d5badd66498905f300e0413 Nov 27 07:21:12 crc kubenswrapper[4706]: I1127 07:21:12.365960 4706 generic.go:334] "Generic (PLEG): container finished" podID="94169577-2199-446a-9d53-8473269fc045" containerID="3eb746183d3299d176023504f7c1a7ecfdfccb4456c29386fad43657c43cd0fa" exitCode=0 Nov 27 07:21:12 crc kubenswrapper[4706]: I1127 07:21:12.366250 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" event={"ID":"94169577-2199-446a-9d53-8473269fc045","Type":"ContainerDied","Data":"3eb746183d3299d176023504f7c1a7ecfdfccb4456c29386fad43657c43cd0fa"} Nov 27 07:21:12 crc kubenswrapper[4706]: I1127 07:21:12.366281 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" event={"ID":"94169577-2199-446a-9d53-8473269fc045","Type":"ContainerStarted","Data":"46eb2e1b6147319d52b871094fac9a43a335a7e20d5badd66498905f300e0413"} Nov 27 07:21:14 crc kubenswrapper[4706]: I1127 07:21:14.379355 4706 generic.go:334] "Generic (PLEG): container finished" podID="94169577-2199-446a-9d53-8473269fc045" containerID="090f20d74cd50b5f29e6c1f1d01ea1ccbc390f32108f3bce77c71c2285bcae32" exitCode=0 Nov 27 07:21:14 crc kubenswrapper[4706]: I1127 07:21:14.379408 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" event={"ID":"94169577-2199-446a-9d53-8473269fc045","Type":"ContainerDied","Data":"090f20d74cd50b5f29e6c1f1d01ea1ccbc390f32108f3bce77c71c2285bcae32"} Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.178145 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.178496 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.178543 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.179069 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4cee0d2f37ffae96c2927fba888af9c607b55ddd51e1dd0044435dc5bccfed27"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.179129 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://4cee0d2f37ffae96c2927fba888af9c607b55ddd51e1dd0044435dc5bccfed27" gracePeriod=600 Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.387252 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="4cee0d2f37ffae96c2927fba888af9c607b55ddd51e1dd0044435dc5bccfed27" exitCode=0 Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.387326 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"4cee0d2f37ffae96c2927fba888af9c607b55ddd51e1dd0044435dc5bccfed27"} Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.387378 4706 scope.go:117] "RemoveContainer" containerID="7171128fb93f11a5548129b54f5af0ff478bef072ea546bb401cec5c726024db" Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.389209 4706 generic.go:334] "Generic (PLEG): container finished" podID="94169577-2199-446a-9d53-8473269fc045" containerID="f2cbe649faeed115dff4c5e03f81d485700d3dfd7addd47ff9fdb7a522a5b977" exitCode=0 Nov 27 07:21:15 crc kubenswrapper[4706]: I1127 07:21:15.389259 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" event={"ID":"94169577-2199-446a-9d53-8473269fc045","Type":"ContainerDied","Data":"f2cbe649faeed115dff4c5e03f81d485700d3dfd7addd47ff9fdb7a522a5b977"} Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.401082 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"db4ade098e79285766f3d1f1fb101ceeccbe33f1280ddb92a2252c3b8d5a2939"} Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.669936 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.802096 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-bundle\") pod \"94169577-2199-446a-9d53-8473269fc045\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.802137 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-util\") pod \"94169577-2199-446a-9d53-8473269fc045\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.802214 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q47dq\" (UniqueName: \"kubernetes.io/projected/94169577-2199-446a-9d53-8473269fc045-kube-api-access-q47dq\") pod \"94169577-2199-446a-9d53-8473269fc045\" (UID: \"94169577-2199-446a-9d53-8473269fc045\") " Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.804925 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-bundle" (OuterVolumeSpecName: "bundle") pod "94169577-2199-446a-9d53-8473269fc045" (UID: "94169577-2199-446a-9d53-8473269fc045"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.807573 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94169577-2199-446a-9d53-8473269fc045-kube-api-access-q47dq" (OuterVolumeSpecName: "kube-api-access-q47dq") pod "94169577-2199-446a-9d53-8473269fc045" (UID: "94169577-2199-446a-9d53-8473269fc045"). InnerVolumeSpecName "kube-api-access-q47dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.904408 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.904462 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q47dq\" (UniqueName: \"kubernetes.io/projected/94169577-2199-446a-9d53-8473269fc045-kube-api-access-q47dq\") on node \"crc\" DevicePath \"\"" Nov 27 07:21:16 crc kubenswrapper[4706]: I1127 07:21:16.954193 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-util" (OuterVolumeSpecName: "util") pod "94169577-2199-446a-9d53-8473269fc045" (UID: "94169577-2199-446a-9d53-8473269fc045"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:21:17 crc kubenswrapper[4706]: I1127 07:21:17.005269 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94169577-2199-446a-9d53-8473269fc045-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:21:17 crc kubenswrapper[4706]: I1127 07:21:17.408439 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" event={"ID":"94169577-2199-446a-9d53-8473269fc045","Type":"ContainerDied","Data":"46eb2e1b6147319d52b871094fac9a43a335a7e20d5badd66498905f300e0413"} Nov 27 07:21:17 crc kubenswrapper[4706]: I1127 07:21:17.408845 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46eb2e1b6147319d52b871094fac9a43a335a7e20d5badd66498905f300e0413" Nov 27 07:21:17 crc kubenswrapper[4706]: I1127 07:21:17.408459 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.308382 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp"] Nov 27 07:21:22 crc kubenswrapper[4706]: E1127 07:21:22.309170 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94169577-2199-446a-9d53-8473269fc045" containerName="util" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.309182 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="94169577-2199-446a-9d53-8473269fc045" containerName="util" Nov 27 07:21:22 crc kubenswrapper[4706]: E1127 07:21:22.309193 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94169577-2199-446a-9d53-8473269fc045" containerName="extract" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.309200 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="94169577-2199-446a-9d53-8473269fc045" containerName="extract" Nov 27 07:21:22 crc kubenswrapper[4706]: E1127 07:21:22.309210 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94169577-2199-446a-9d53-8473269fc045" containerName="pull" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.309216 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="94169577-2199-446a-9d53-8473269fc045" containerName="pull" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.309337 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="94169577-2199-446a-9d53-8473269fc045" containerName="extract" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.309902 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.312086 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.312262 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nkgcz" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.323411 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp"] Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.470898 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-webhook-cert\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.471020 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv894\" (UniqueName: \"kubernetes.io/projected/006027de-9ecc-4ff3-96a6-6918afb24789-kube-api-access-dv894\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.471078 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-apiservice-cert\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.572460 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv894\" (UniqueName: \"kubernetes.io/projected/006027de-9ecc-4ff3-96a6-6918afb24789-kube-api-access-dv894\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.572531 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-apiservice-cert\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.572559 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-webhook-cert\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.578285 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-apiservice-cert\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.578388 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-webhook-cert\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.593876 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv894\" (UniqueName: \"kubernetes.io/projected/006027de-9ecc-4ff3-96a6-6918afb24789-kube-api-access-dv894\") pod \"infra-operator-controller-manager-79755fc49-rb6vp\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:22 crc kubenswrapper[4706]: I1127 07:21:22.628867 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:23 crc kubenswrapper[4706]: I1127 07:21:23.074470 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp"] Nov 27 07:21:23 crc kubenswrapper[4706]: I1127 07:21:23.440603 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" event={"ID":"006027de-9ecc-4ff3-96a6-6918afb24789","Type":"ContainerStarted","Data":"d8d749dc5c2a268095ace272046c549d3b0231a1c95164d04b21b67440e88aae"} Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.454436 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" event={"ID":"006027de-9ecc-4ff3-96a6-6918afb24789","Type":"ContainerStarted","Data":"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c"} Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.715522 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.717063 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.719580 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.720080 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-8kwnw" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.720240 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.720385 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.723629 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.724658 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.726167 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.730207 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.732448 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.740470 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.749342 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.760320 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918144 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/170c6f78-b315-483f-a997-bfed203bacb7-config-data-generated\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918181 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-operator-scripts\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918235 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kolla-config\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918258 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzlr9\" (UniqueName: \"kubernetes.io/projected/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kube-api-access-jzlr9\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918321 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-kolla-config\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918336 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3047c710-a2e6-455a-951a-88971379f86f-config-data-generated\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918355 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-config-data-default\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918371 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-kolla-config\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918398 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918412 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72dzc\" (UniqueName: \"kubernetes.io/projected/3047c710-a2e6-455a-951a-88971379f86f-kube-api-access-72dzc\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918430 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-operator-scripts\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918471 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918496 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-config-data-default\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918514 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918531 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918580 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-default\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918598 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzj8b\" (UniqueName: \"kubernetes.io/projected/170c6f78-b315-483f-a997-bfed203bacb7-kube-api-access-pzj8b\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:25 crc kubenswrapper[4706]: I1127 07:21:25.918616 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019711 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/170c6f78-b315-483f-a997-bfed203bacb7-config-data-generated\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019762 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-operator-scripts\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019799 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kolla-config\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019831 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzlr9\" (UniqueName: \"kubernetes.io/projected/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kube-api-access-jzlr9\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019852 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-kolla-config\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019868 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3047c710-a2e6-455a-951a-88971379f86f-config-data-generated\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019882 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-config-data-default\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-kolla-config\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019936 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019953 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72dzc\" (UniqueName: \"kubernetes.io/projected/3047c710-a2e6-455a-951a-88971379f86f-kube-api-access-72dzc\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.019971 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-operator-scripts\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.020030 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.020048 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-config-data-default\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.020360 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.020375 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-config-data-default\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021101 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-kolla-config\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021112 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-config-data-default\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021211 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021340 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kolla-config\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021392 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021428 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021447 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-default\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021464 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzj8b\" (UniqueName: \"kubernetes.io/projected/170c6f78-b315-483f-a997-bfed203bacb7-kube-api-access-pzj8b\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021760 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.022515 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3047c710-a2e6-455a-951a-88971379f86f-config-data-generated\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.022539 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.022359 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-operator-scripts\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.022367 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-default\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.021971 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-operator-scripts\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.023077 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.023201 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-kolla-config\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.028390 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/170c6f78-b315-483f-a997-bfed203bacb7-config-data-generated\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.042363 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.044695 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.045061 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.046008 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzlr9\" (UniqueName: \"kubernetes.io/projected/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kube-api-access-jzlr9\") pod \"openstack-galera-0\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.047154 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72dzc\" (UniqueName: \"kubernetes.io/projected/3047c710-a2e6-455a-951a-88971379f86f-kube-api-access-72dzc\") pod \"openstack-galera-1\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.049435 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzj8b\" (UniqueName: \"kubernetes.io/projected/170c6f78-b315-483f-a997-bfed203bacb7-kube-api-access-pzj8b\") pod \"openstack-galera-2\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.057531 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.072097 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.345753 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.486260 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 27 07:21:26 crc kubenswrapper[4706]: W1127 07:21:26.519115 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod170c6f78_b315_483f_a997_bfed203bacb7.slice/crio-f6efe971654605f901f8eccb5ba3479ed48fdbb7294614313744438cdef4c6a1 WatchSource:0}: Error finding container f6efe971654605f901f8eccb5ba3479ed48fdbb7294614313744438cdef4c6a1: Status 404 returned error can't find the container with id f6efe971654605f901f8eccb5ba3479ed48fdbb7294614313744438cdef4c6a1 Nov 27 07:21:26 crc kubenswrapper[4706]: I1127 07:21:26.738185 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 27 07:21:27 crc kubenswrapper[4706]: W1127 07:21:27.032791 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3047c710_a2e6_455a_951a_88971379f86f.slice/crio-1f839b1f75a6d582efeaa36446071fb0d0082f8612949d9c304e2d7de4939369 WatchSource:0}: Error finding container 1f839b1f75a6d582efeaa36446071fb0d0082f8612949d9c304e2d7de4939369: Status 404 returned error can't find the container with id 1f839b1f75a6d582efeaa36446071fb0d0082f8612949d9c304e2d7de4939369 Nov 27 07:21:27 crc kubenswrapper[4706]: I1127 07:21:27.421282 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 27 07:21:27 crc kubenswrapper[4706]: I1127 07:21:27.470046 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" event={"ID":"006027de-9ecc-4ff3-96a6-6918afb24789","Type":"ContainerStarted","Data":"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9"} Nov 27 07:21:27 crc kubenswrapper[4706]: I1127 07:21:27.470192 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:27 crc kubenswrapper[4706]: I1127 07:21:27.471689 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3047c710-a2e6-455a-951a-88971379f86f","Type":"ContainerStarted","Data":"1f839b1f75a6d582efeaa36446071fb0d0082f8612949d9c304e2d7de4939369"} Nov 27 07:21:27 crc kubenswrapper[4706]: I1127 07:21:27.472906 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d","Type":"ContainerStarted","Data":"7d8993f378a9ef0c8619418490c1b47b3d36cf1cab0b7ac7d1fd785962f65c80"} Nov 27 07:21:27 crc kubenswrapper[4706]: I1127 07:21:27.474108 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"170c6f78-b315-483f-a997-bfed203bacb7","Type":"ContainerStarted","Data":"f6efe971654605f901f8eccb5ba3479ed48fdbb7294614313744438cdef4c6a1"} Nov 27 07:21:27 crc kubenswrapper[4706]: I1127 07:21:27.495014 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" podStartSLOduration=1.505581139 podStartE2EDuration="5.494995771s" podCreationTimestamp="2025-11-27 07:21:22 +0000 UTC" firstStartedPulling="2025-11-27 07:21:23.079601082 +0000 UTC m=+766.969191892" lastFinishedPulling="2025-11-27 07:21:27.069015714 +0000 UTC m=+770.958606524" observedRunningTime="2025-11-27 07:21:27.489712211 +0000 UTC m=+771.379303031" watchObservedRunningTime="2025-11-27 07:21:27.494995771 +0000 UTC m=+771.384586581" Nov 27 07:21:32 crc kubenswrapper[4706]: I1127 07:21:32.632739 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:21:34 crc kubenswrapper[4706]: I1127 07:21:34.521275 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3047c710-a2e6-455a-951a-88971379f86f","Type":"ContainerStarted","Data":"65df9453f65098dc10938c2788c8a4b222dd248676351195b7bf545aec7cdf6e"} Nov 27 07:21:34 crc kubenswrapper[4706]: I1127 07:21:34.523612 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d","Type":"ContainerStarted","Data":"7340b04b532e0329384704ffac9894d728fbe34a4cc96e8282e264d9a7a5e182"} Nov 27 07:21:34 crc kubenswrapper[4706]: I1127 07:21:34.525144 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"170c6f78-b315-483f-a997-bfed203bacb7","Type":"ContainerStarted","Data":"83d37dbcc8861c7c72d29999f0b93082cab4a7ed666c6efadb361064699d6e49"} Nov 27 07:21:36 crc kubenswrapper[4706]: I1127 07:21:36.900754 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 27 07:21:36 crc kubenswrapper[4706]: I1127 07:21:36.902558 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:36 crc kubenswrapper[4706]: I1127 07:21:36.904192 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-k4rlh" Nov 27 07:21:36 crc kubenswrapper[4706]: I1127 07:21:36.907725 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 27 07:21:36 crc kubenswrapper[4706]: I1127 07:21:36.912374 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.000047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kolla-config\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.000115 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7kkb\" (UniqueName: \"kubernetes.io/projected/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kube-api-access-l7kkb\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.000143 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-config-data\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.100742 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-config-data\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.100834 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kolla-config\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.100867 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7kkb\" (UniqueName: \"kubernetes.io/projected/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kube-api-access-l7kkb\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.101722 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-config-data\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.101722 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kolla-config\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.144290 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7kkb\" (UniqueName: \"kubernetes.io/projected/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kube-api-access-l7kkb\") pod \"memcached-0\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.227413 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:37 crc kubenswrapper[4706]: I1127 07:21:37.599751 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 27 07:21:38 crc kubenswrapper[4706]: I1127 07:21:38.558740 4706 generic.go:334] "Generic (PLEG): container finished" podID="170c6f78-b315-483f-a997-bfed203bacb7" containerID="83d37dbcc8861c7c72d29999f0b93082cab4a7ed666c6efadb361064699d6e49" exitCode=0 Nov 27 07:21:38 crc kubenswrapper[4706]: I1127 07:21:38.559888 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"170c6f78-b315-483f-a997-bfed203bacb7","Type":"ContainerDied","Data":"83d37dbcc8861c7c72d29999f0b93082cab4a7ed666c6efadb361064699d6e49"} Nov 27 07:21:38 crc kubenswrapper[4706]: I1127 07:21:38.559987 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51","Type":"ContainerStarted","Data":"9acfd22bd8ce2b32c15aaa5c7bfca9811fb708e043b816efdc02295ab7b683f0"} Nov 27 07:21:38 crc kubenswrapper[4706]: I1127 07:21:38.561790 4706 generic.go:334] "Generic (PLEG): container finished" podID="3047c710-a2e6-455a-951a-88971379f86f" containerID="65df9453f65098dc10938c2788c8a4b222dd248676351195b7bf545aec7cdf6e" exitCode=0 Nov 27 07:21:38 crc kubenswrapper[4706]: I1127 07:21:38.561830 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3047c710-a2e6-455a-951a-88971379f86f","Type":"ContainerDied","Data":"65df9453f65098dc10938c2788c8a4b222dd248676351195b7bf545aec7cdf6e"} Nov 27 07:21:38 crc kubenswrapper[4706]: I1127 07:21:38.569378 4706 generic.go:334] "Generic (PLEG): container finished" podID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerID="7340b04b532e0329384704ffac9894d728fbe34a4cc96e8282e264d9a7a5e182" exitCode=0 Nov 27 07:21:38 crc kubenswrapper[4706]: I1127 07:21:38.569418 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d","Type":"ContainerDied","Data":"7340b04b532e0329384704ffac9894d728fbe34a4cc96e8282e264d9a7a5e182"} Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.578365 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3047c710-a2e6-455a-951a-88971379f86f","Type":"ContainerStarted","Data":"db924923123328745d70c4e3e07509a6c0a28b588ec32e6d8cd509d486ad926b"} Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.580858 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d","Type":"ContainerStarted","Data":"d1104eea8593de00b4dfb366253def29206277bf5e1f33d94f3c0712811050fa"} Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.586071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"170c6f78-b315-483f-a997-bfed203bacb7","Type":"ContainerStarted","Data":"9a73ed5960d9161255e97432f54ed2f8b667401423d038d0f7ed0aca9c65379d"} Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.607573 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=8.556148339 podStartE2EDuration="15.607558134s" podCreationTimestamp="2025-11-27 07:21:24 +0000 UTC" firstStartedPulling="2025-11-27 07:21:27.036039153 +0000 UTC m=+770.925629963" lastFinishedPulling="2025-11-27 07:21:34.087448948 +0000 UTC m=+777.977039758" observedRunningTime="2025-11-27 07:21:39.604741055 +0000 UTC m=+783.494331885" watchObservedRunningTime="2025-11-27 07:21:39.607558134 +0000 UTC m=+783.497148944" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.621916 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=8.992538232 podStartE2EDuration="15.621896337s" podCreationTimestamp="2025-11-27 07:21:24 +0000 UTC" firstStartedPulling="2025-11-27 07:21:27.431236389 +0000 UTC m=+771.320827199" lastFinishedPulling="2025-11-27 07:21:34.060594474 +0000 UTC m=+777.950185304" observedRunningTime="2025-11-27 07:21:39.621885877 +0000 UTC m=+783.511476687" watchObservedRunningTime="2025-11-27 07:21:39.621896337 +0000 UTC m=+783.511487147" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.645739 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=8.115879012 podStartE2EDuration="15.64572382s" podCreationTimestamp="2025-11-27 07:21:24 +0000 UTC" firstStartedPulling="2025-11-27 07:21:26.521660693 +0000 UTC m=+770.411251503" lastFinishedPulling="2025-11-27 07:21:34.051505501 +0000 UTC m=+777.941096311" observedRunningTime="2025-11-27 07:21:39.643114148 +0000 UTC m=+783.532704958" watchObservedRunningTime="2025-11-27 07:21:39.64572382 +0000 UTC m=+783.535314630" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.679077 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-bh7ff"] Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.679776 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.684004 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-4hj4f" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.689888 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-bh7ff"] Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.862213 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngzqj\" (UniqueName: \"kubernetes.io/projected/c1a79656-0142-4194-bd13-90928f67a27e-kube-api-access-ngzqj\") pod \"rabbitmq-cluster-operator-index-bh7ff\" (UID: \"c1a79656-0142-4194-bd13-90928f67a27e\") " pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.963586 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngzqj\" (UniqueName: \"kubernetes.io/projected/c1a79656-0142-4194-bd13-90928f67a27e-kube-api-access-ngzqj\") pod \"rabbitmq-cluster-operator-index-bh7ff\" (UID: \"c1a79656-0142-4194-bd13-90928f67a27e\") " pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.981205 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngzqj\" (UniqueName: \"kubernetes.io/projected/c1a79656-0142-4194-bd13-90928f67a27e-kube-api-access-ngzqj\") pod \"rabbitmq-cluster-operator-index-bh7ff\" (UID: \"c1a79656-0142-4194-bd13-90928f67a27e\") " pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:39 crc kubenswrapper[4706]: I1127 07:21:39.993367 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:40 crc kubenswrapper[4706]: I1127 07:21:40.592778 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51","Type":"ContainerStarted","Data":"7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6"} Nov 27 07:21:40 crc kubenswrapper[4706]: I1127 07:21:40.593102 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:40 crc kubenswrapper[4706]: I1127 07:21:40.612513 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=1.827586874 podStartE2EDuration="4.612494474s" podCreationTimestamp="2025-11-27 07:21:36 +0000 UTC" firstStartedPulling="2025-11-27 07:21:37.619504963 +0000 UTC m=+781.509095773" lastFinishedPulling="2025-11-27 07:21:40.404412563 +0000 UTC m=+784.294003373" observedRunningTime="2025-11-27 07:21:40.611560825 +0000 UTC m=+784.501151635" watchObservedRunningTime="2025-11-27 07:21:40.612494474 +0000 UTC m=+784.502085284" Nov 27 07:21:40 crc kubenswrapper[4706]: I1127 07:21:40.760214 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-bh7ff"] Nov 27 07:21:40 crc kubenswrapper[4706]: W1127 07:21:40.771746 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1a79656_0142_4194_bd13_90928f67a27e.slice/crio-c22969ec6cabf22da245f81038dacd1d4d694d3474fc331bdd48901e16a425f1 WatchSource:0}: Error finding container c22969ec6cabf22da245f81038dacd1d4d694d3474fc331bdd48901e16a425f1: Status 404 returned error can't find the container with id c22969ec6cabf22da245f81038dacd1d4d694d3474fc331bdd48901e16a425f1 Nov 27 07:21:41 crc kubenswrapper[4706]: I1127 07:21:41.600073 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" event={"ID":"c1a79656-0142-4194-bd13-90928f67a27e","Type":"ContainerStarted","Data":"c22969ec6cabf22da245f81038dacd1d4d694d3474fc331bdd48901e16a425f1"} Nov 27 07:21:44 crc kubenswrapper[4706]: I1127 07:21:44.621775 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" event={"ID":"c1a79656-0142-4194-bd13-90928f67a27e","Type":"ContainerStarted","Data":"f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b"} Nov 27 07:21:44 crc kubenswrapper[4706]: I1127 07:21:44.643619 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" podStartSLOduration=2.645732648 podStartE2EDuration="5.643599142s" podCreationTimestamp="2025-11-27 07:21:39 +0000 UTC" firstStartedPulling="2025-11-27 07:21:40.773266003 +0000 UTC m=+784.662856823" lastFinishedPulling="2025-11-27 07:21:43.771132507 +0000 UTC m=+787.660723317" observedRunningTime="2025-11-27 07:21:44.638366847 +0000 UTC m=+788.527957687" watchObservedRunningTime="2025-11-27 07:21:44.643599142 +0000 UTC m=+788.533189982" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.058426 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.058750 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.072683 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.072748 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.189830 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.346281 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.346436 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:46 crc kubenswrapper[4706]: I1127 07:21:46.727261 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:21:47 crc kubenswrapper[4706]: I1127 07:21:47.229157 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 27 07:21:49 crc kubenswrapper[4706]: I1127 07:21:49.994277 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:49 crc kubenswrapper[4706]: I1127 07:21:49.994516 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:50 crc kubenswrapper[4706]: I1127 07:21:50.027763 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:50 crc kubenswrapper[4706]: I1127 07:21:50.680920 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.301535 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx"] Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.302932 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.304536 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dk9r5" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.315706 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx"] Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.437466 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnpwv\" (UniqueName: \"kubernetes.io/projected/0371adf5-a190-4d98-90db-b83dcd65b5b9-kube-api-access-tnpwv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.437539 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.437564 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.538711 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnpwv\" (UniqueName: \"kubernetes.io/projected/0371adf5-a190-4d98-90db-b83dcd65b5b9-kube-api-access-tnpwv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.538802 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.538839 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.539304 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.539601 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.559830 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnpwv\" (UniqueName: \"kubernetes.io/projected/0371adf5-a190-4d98-90db-b83dcd65b5b9-kube-api-access-tnpwv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:52 crc kubenswrapper[4706]: I1127 07:21:52.617951 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:53 crc kubenswrapper[4706]: I1127 07:21:53.051106 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx"] Nov 27 07:21:53 crc kubenswrapper[4706]: I1127 07:21:53.679303 4706 generic.go:334] "Generic (PLEG): container finished" podID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerID="a5ea7b183ead5b7612c935973b11381149d3cd0f87a35c1ea1ba8cec6d2c884d" exitCode=0 Nov 27 07:21:53 crc kubenswrapper[4706]: I1127 07:21:53.679366 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" event={"ID":"0371adf5-a190-4d98-90db-b83dcd65b5b9","Type":"ContainerDied","Data":"a5ea7b183ead5b7612c935973b11381149d3cd0f87a35c1ea1ba8cec6d2c884d"} Nov 27 07:21:53 crc kubenswrapper[4706]: I1127 07:21:53.679436 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" event={"ID":"0371adf5-a190-4d98-90db-b83dcd65b5b9","Type":"ContainerStarted","Data":"e137e70eada96709aeebd93dd3ce3ef797f94f07746c4c8dc1e01daa10c29fdf"} Nov 27 07:21:54 crc kubenswrapper[4706]: I1127 07:21:54.687350 4706 generic.go:334] "Generic (PLEG): container finished" podID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerID="8e3901e1155389115253f4ec48b1d9d325f803566fd819e9b1ceb383df2d3d3b" exitCode=0 Nov 27 07:21:54 crc kubenswrapper[4706]: I1127 07:21:54.687437 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" event={"ID":"0371adf5-a190-4d98-90db-b83dcd65b5b9","Type":"ContainerDied","Data":"8e3901e1155389115253f4ec48b1d9d325f803566fd819e9b1ceb383df2d3d3b"} Nov 27 07:21:55 crc kubenswrapper[4706]: I1127 07:21:55.695403 4706 generic.go:334] "Generic (PLEG): container finished" podID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerID="fba8bc694391c53a203d03f959f74bbd5575b1c342b6a2a1e47d4435cba08038" exitCode=0 Nov 27 07:21:55 crc kubenswrapper[4706]: I1127 07:21:55.695458 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" event={"ID":"0371adf5-a190-4d98-90db-b83dcd65b5b9","Type":"ContainerDied","Data":"fba8bc694391c53a203d03f959f74bbd5575b1c342b6a2a1e47d4435cba08038"} Nov 27 07:21:56 crc kubenswrapper[4706]: I1127 07:21:56.156904 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="170c6f78-b315-483f-a997-bfed203bacb7" containerName="galera" probeResult="failure" output=< Nov 27 07:21:56 crc kubenswrapper[4706]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Nov 27 07:21:56 crc kubenswrapper[4706]: > Nov 27 07:21:56 crc kubenswrapper[4706]: I1127 07:21:56.247409 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:56 crc kubenswrapper[4706]: I1127 07:21:56.311554 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:21:56 crc kubenswrapper[4706]: I1127 07:21:56.975752 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.102950 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-bundle\") pod \"0371adf5-a190-4d98-90db-b83dcd65b5b9\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.103012 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-util\") pod \"0371adf5-a190-4d98-90db-b83dcd65b5b9\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.103071 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnpwv\" (UniqueName: \"kubernetes.io/projected/0371adf5-a190-4d98-90db-b83dcd65b5b9-kube-api-access-tnpwv\") pod \"0371adf5-a190-4d98-90db-b83dcd65b5b9\" (UID: \"0371adf5-a190-4d98-90db-b83dcd65b5b9\") " Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.103647 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-bundle" (OuterVolumeSpecName: "bundle") pod "0371adf5-a190-4d98-90db-b83dcd65b5b9" (UID: "0371adf5-a190-4d98-90db-b83dcd65b5b9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.108423 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0371adf5-a190-4d98-90db-b83dcd65b5b9-kube-api-access-tnpwv" (OuterVolumeSpecName: "kube-api-access-tnpwv") pod "0371adf5-a190-4d98-90db-b83dcd65b5b9" (UID: "0371adf5-a190-4d98-90db-b83dcd65b5b9"). InnerVolumeSpecName "kube-api-access-tnpwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.117420 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-util" (OuterVolumeSpecName: "util") pod "0371adf5-a190-4d98-90db-b83dcd65b5b9" (UID: "0371adf5-a190-4d98-90db-b83dcd65b5b9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.205138 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.205178 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnpwv\" (UniqueName: \"kubernetes.io/projected/0371adf5-a190-4d98-90db-b83dcd65b5b9-kube-api-access-tnpwv\") on node \"crc\" DevicePath \"\"" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.205196 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0371adf5-a190-4d98-90db-b83dcd65b5b9-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.710744 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" event={"ID":"0371adf5-a190-4d98-90db-b83dcd65b5b9","Type":"ContainerDied","Data":"e137e70eada96709aeebd93dd3ce3ef797f94f07746c4c8dc1e01daa10c29fdf"} Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.710783 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e137e70eada96709aeebd93dd3ce3ef797f94f07746c4c8dc1e01daa10c29fdf" Nov 27 07:21:57 crc kubenswrapper[4706]: I1127 07:21:57.710798 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx" Nov 27 07:22:00 crc kubenswrapper[4706]: I1127 07:22:00.659693 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:22:00 crc kubenswrapper[4706]: I1127 07:22:00.730026 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.927206 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb"] Nov 27 07:22:06 crc kubenswrapper[4706]: E1127 07:22:06.928153 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerName="extract" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.928168 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerName="extract" Nov 27 07:22:06 crc kubenswrapper[4706]: E1127 07:22:06.928191 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerName="util" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.928199 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerName="util" Nov 27 07:22:06 crc kubenswrapper[4706]: E1127 07:22:06.928213 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerName="pull" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.928240 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerName="pull" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.928375 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" containerName="extract" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.928912 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.931131 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-snrbf" Nov 27 07:22:06 crc kubenswrapper[4706]: I1127 07:22:06.953043 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb"] Nov 27 07:22:07 crc kubenswrapper[4706]: I1127 07:22:07.058166 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgljh\" (UniqueName: \"kubernetes.io/projected/636f33ce-c9d8-46ce-99e3-db7408c6e65b-kube-api-access-dgljh\") pod \"rabbitmq-cluster-operator-779fc9694b-px4kb\" (UID: \"636f33ce-c9d8-46ce-99e3-db7408c6e65b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" Nov 27 07:22:07 crc kubenswrapper[4706]: I1127 07:22:07.159416 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgljh\" (UniqueName: \"kubernetes.io/projected/636f33ce-c9d8-46ce-99e3-db7408c6e65b-kube-api-access-dgljh\") pod \"rabbitmq-cluster-operator-779fc9694b-px4kb\" (UID: \"636f33ce-c9d8-46ce-99e3-db7408c6e65b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" Nov 27 07:22:07 crc kubenswrapper[4706]: I1127 07:22:07.177687 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgljh\" (UniqueName: \"kubernetes.io/projected/636f33ce-c9d8-46ce-99e3-db7408c6e65b-kube-api-access-dgljh\") pod \"rabbitmq-cluster-operator-779fc9694b-px4kb\" (UID: \"636f33ce-c9d8-46ce-99e3-db7408c6e65b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" Nov 27 07:22:07 crc kubenswrapper[4706]: I1127 07:22:07.253486 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" Nov 27 07:22:07 crc kubenswrapper[4706]: I1127 07:22:07.654788 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb"] Nov 27 07:22:07 crc kubenswrapper[4706]: W1127 07:22:07.658097 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod636f33ce_c9d8_46ce_99e3_db7408c6e65b.slice/crio-373e22adbc8ce2bfd7c5abaf2e21213401642df17d0bcbb2889b9b7e93589168 WatchSource:0}: Error finding container 373e22adbc8ce2bfd7c5abaf2e21213401642df17d0bcbb2889b9b7e93589168: Status 404 returned error can't find the container with id 373e22adbc8ce2bfd7c5abaf2e21213401642df17d0bcbb2889b9b7e93589168 Nov 27 07:22:07 crc kubenswrapper[4706]: I1127 07:22:07.773727 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" event={"ID":"636f33ce-c9d8-46ce-99e3-db7408c6e65b","Type":"ContainerStarted","Data":"373e22adbc8ce2bfd7c5abaf2e21213401642df17d0bcbb2889b9b7e93589168"} Nov 27 07:22:10 crc kubenswrapper[4706]: I1127 07:22:10.807986 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" event={"ID":"636f33ce-c9d8-46ce-99e3-db7408c6e65b","Type":"ContainerStarted","Data":"b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844"} Nov 27 07:22:10 crc kubenswrapper[4706]: I1127 07:22:10.821549 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" podStartSLOduration=2.047870025 podStartE2EDuration="4.821530728s" podCreationTimestamp="2025-11-27 07:22:06 +0000 UTC" firstStartedPulling="2025-11-27 07:22:07.660089377 +0000 UTC m=+811.549680187" lastFinishedPulling="2025-11-27 07:22:10.43375008 +0000 UTC m=+814.323340890" observedRunningTime="2025-11-27 07:22:10.821516878 +0000 UTC m=+814.711107688" watchObservedRunningTime="2025-11-27 07:22:10.821530728 +0000 UTC m=+814.711121548" Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.910442 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.913159 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.915183 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.915652 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-2k5r4" Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.916049 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.916192 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.916493 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 27 07:22:13 crc kubenswrapper[4706]: I1127 07:22:13.929027 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052153 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052249 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28glv\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-kube-api-access-28glv\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052271 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53a204c4-9b3c-4d83-a87f-bf48abe46f77-pod-info\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052303 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53a204c4-9b3c-4d83-a87f-bf48abe46f77-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052318 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052540 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53a204c4-9b3c-4d83-a87f-bf48abe46f77-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.052665 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.154507 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53a204c4-9b3c-4d83-a87f-bf48abe46f77-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.154703 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.154785 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.154838 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.154925 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28glv\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-kube-api-access-28glv\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.154992 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53a204c4-9b3c-4d83-a87f-bf48abe46f77-pod-info\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.155579 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53a204c4-9b3c-4d83-a87f-bf48abe46f77-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.155654 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.156144 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.156177 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.156606 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53a204c4-9b3c-4d83-a87f-bf48abe46f77-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.161185 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53a204c4-9b3c-4d83-a87f-bf48abe46f77-pod-info\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.169814 4706 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.170115 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2a5420bac897cbbbe797ed2d728ef7b01fb57ad668146808505879aa38cf1949/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.171899 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53a204c4-9b3c-4d83-a87f-bf48abe46f77-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.175787 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.179569 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28glv\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-kube-api-access-28glv\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.218433 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\") pod \"rabbitmq-server-0\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.231837 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.682078 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 27 07:22:14 crc kubenswrapper[4706]: I1127 07:22:14.837082 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"53a204c4-9b3c-4d83-a87f-bf48abe46f77","Type":"ContainerStarted","Data":"655d8fb3c393e0e5cdb0f05d5e3d64cf0a6b4496c84295ecd358435cd6ce6715"} Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.467835 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-5slwj"] Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.468888 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.471257 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-5w6q8" Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.477717 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-5slwj"] Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.578490 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx72s\" (UniqueName: \"kubernetes.io/projected/82a85ca2-e74d-48ef-8fb4-19edc42cddac-kube-api-access-tx72s\") pod \"keystone-operator-index-5slwj\" (UID: \"82a85ca2-e74d-48ef-8fb4-19edc42cddac\") " pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.680447 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx72s\" (UniqueName: \"kubernetes.io/projected/82a85ca2-e74d-48ef-8fb4-19edc42cddac-kube-api-access-tx72s\") pod \"keystone-operator-index-5slwj\" (UID: \"82a85ca2-e74d-48ef-8fb4-19edc42cddac\") " pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.720296 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx72s\" (UniqueName: \"kubernetes.io/projected/82a85ca2-e74d-48ef-8fb4-19edc42cddac-kube-api-access-tx72s\") pod \"keystone-operator-index-5slwj\" (UID: \"82a85ca2-e74d-48ef-8fb4-19edc42cddac\") " pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:15 crc kubenswrapper[4706]: I1127 07:22:15.798411 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:16 crc kubenswrapper[4706]: I1127 07:22:16.238724 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-5slwj"] Nov 27 07:22:16 crc kubenswrapper[4706]: W1127 07:22:16.261602 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82a85ca2_e74d_48ef_8fb4_19edc42cddac.slice/crio-4074a2934ec3e43ae415d3f2165177e1e455a2e81cf71cb49c069d38dd866cae WatchSource:0}: Error finding container 4074a2934ec3e43ae415d3f2165177e1e455a2e81cf71cb49c069d38dd866cae: Status 404 returned error can't find the container with id 4074a2934ec3e43ae415d3f2165177e1e455a2e81cf71cb49c069d38dd866cae Nov 27 07:22:16 crc kubenswrapper[4706]: I1127 07:22:16.853269 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5slwj" event={"ID":"82a85ca2-e74d-48ef-8fb4-19edc42cddac","Type":"ContainerStarted","Data":"4074a2934ec3e43ae415d3f2165177e1e455a2e81cf71cb49c069d38dd866cae"} Nov 27 07:22:20 crc kubenswrapper[4706]: I1127 07:22:20.885243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"53a204c4-9b3c-4d83-a87f-bf48abe46f77","Type":"ContainerStarted","Data":"a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f"} Nov 27 07:22:20 crc kubenswrapper[4706]: I1127 07:22:20.886817 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5slwj" event={"ID":"82a85ca2-e74d-48ef-8fb4-19edc42cddac","Type":"ContainerStarted","Data":"57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573"} Nov 27 07:22:25 crc kubenswrapper[4706]: I1127 07:22:25.799481 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:25 crc kubenswrapper[4706]: I1127 07:22:25.800084 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:25 crc kubenswrapper[4706]: I1127 07:22:25.821932 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:25 crc kubenswrapper[4706]: I1127 07:22:25.839307 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-5slwj" podStartSLOduration=7.029650846 podStartE2EDuration="10.83928683s" podCreationTimestamp="2025-11-27 07:22:15 +0000 UTC" firstStartedPulling="2025-11-27 07:22:16.264270532 +0000 UTC m=+820.153861342" lastFinishedPulling="2025-11-27 07:22:20.073906516 +0000 UTC m=+823.963497326" observedRunningTime="2025-11-27 07:22:20.960457707 +0000 UTC m=+824.850048557" watchObservedRunningTime="2025-11-27 07:22:25.83928683 +0000 UTC m=+829.728877660" Nov 27 07:22:25 crc kubenswrapper[4706]: I1127 07:22:25.949182 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.497273 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx"] Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.498682 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.500938 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dk9r5" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.510068 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx"] Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.632922 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.633009 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k225\" (UniqueName: \"kubernetes.io/projected/02868e73-1352-47bd-8fb4-8697bb9e2587-kube-api-access-7k225\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.633059 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.734957 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k225\" (UniqueName: \"kubernetes.io/projected/02868e73-1352-47bd-8fb4-8697bb9e2587-kube-api-access-7k225\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.735379 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.735707 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.736342 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.736394 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.767481 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k225\" (UniqueName: \"kubernetes.io/projected/02868e73-1352-47bd-8fb4-8697bb9e2587-kube-api-access-7k225\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:28 crc kubenswrapper[4706]: I1127 07:22:28.824866 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:29 crc kubenswrapper[4706]: I1127 07:22:29.038523 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx"] Nov 27 07:22:29 crc kubenswrapper[4706]: W1127 07:22:29.043869 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02868e73_1352_47bd_8fb4_8697bb9e2587.slice/crio-54306a8cbe063b100fd86d1a7dfa546e70d8abdbb94657ab3ff1ae2edbd829f1 WatchSource:0}: Error finding container 54306a8cbe063b100fd86d1a7dfa546e70d8abdbb94657ab3ff1ae2edbd829f1: Status 404 returned error can't find the container with id 54306a8cbe063b100fd86d1a7dfa546e70d8abdbb94657ab3ff1ae2edbd829f1 Nov 27 07:22:29 crc kubenswrapper[4706]: I1127 07:22:29.947061 4706 generic.go:334] "Generic (PLEG): container finished" podID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerID="873eb2e7db7b0b3a8f43e995cf3a2596e3f66e605538ddc07cd169bf4cf99bc3" exitCode=0 Nov 27 07:22:29 crc kubenswrapper[4706]: I1127 07:22:29.947106 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" event={"ID":"02868e73-1352-47bd-8fb4-8697bb9e2587","Type":"ContainerDied","Data":"873eb2e7db7b0b3a8f43e995cf3a2596e3f66e605538ddc07cd169bf4cf99bc3"} Nov 27 07:22:29 crc kubenswrapper[4706]: I1127 07:22:29.947136 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" event={"ID":"02868e73-1352-47bd-8fb4-8697bb9e2587","Type":"ContainerStarted","Data":"54306a8cbe063b100fd86d1a7dfa546e70d8abdbb94657ab3ff1ae2edbd829f1"} Nov 27 07:22:30 crc kubenswrapper[4706]: I1127 07:22:30.954916 4706 generic.go:334] "Generic (PLEG): container finished" podID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerID="9a1d43cb43cde6762ef8a3e95bdd5e4931e6273c991fc9bdd4df634ac90a3056" exitCode=0 Nov 27 07:22:30 crc kubenswrapper[4706]: I1127 07:22:30.954975 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" event={"ID":"02868e73-1352-47bd-8fb4-8697bb9e2587","Type":"ContainerDied","Data":"9a1d43cb43cde6762ef8a3e95bdd5e4931e6273c991fc9bdd4df634ac90a3056"} Nov 27 07:22:31 crc kubenswrapper[4706]: I1127 07:22:31.963211 4706 generic.go:334] "Generic (PLEG): container finished" podID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerID="e2af845e375f59819a9d1bfe3d25bec314670153c8d2c8136ec610faec8e64ef" exitCode=0 Nov 27 07:22:31 crc kubenswrapper[4706]: I1127 07:22:31.963291 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" event={"ID":"02868e73-1352-47bd-8fb4-8697bb9e2587","Type":"ContainerDied","Data":"e2af845e375f59819a9d1bfe3d25bec314670153c8d2c8136ec610faec8e64ef"} Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.305548 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.505205 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-util\") pod \"02868e73-1352-47bd-8fb4-8697bb9e2587\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.505424 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k225\" (UniqueName: \"kubernetes.io/projected/02868e73-1352-47bd-8fb4-8697bb9e2587-kube-api-access-7k225\") pod \"02868e73-1352-47bd-8fb4-8697bb9e2587\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.505556 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-bundle\") pod \"02868e73-1352-47bd-8fb4-8697bb9e2587\" (UID: \"02868e73-1352-47bd-8fb4-8697bb9e2587\") " Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.507062 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-bundle" (OuterVolumeSpecName: "bundle") pod "02868e73-1352-47bd-8fb4-8697bb9e2587" (UID: "02868e73-1352-47bd-8fb4-8697bb9e2587"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.514736 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02868e73-1352-47bd-8fb4-8697bb9e2587-kube-api-access-7k225" (OuterVolumeSpecName: "kube-api-access-7k225") pod "02868e73-1352-47bd-8fb4-8697bb9e2587" (UID: "02868e73-1352-47bd-8fb4-8697bb9e2587"). InnerVolumeSpecName "kube-api-access-7k225". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.522985 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-util" (OuterVolumeSpecName: "util") pod "02868e73-1352-47bd-8fb4-8697bb9e2587" (UID: "02868e73-1352-47bd-8fb4-8697bb9e2587"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.606814 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.606851 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k225\" (UniqueName: \"kubernetes.io/projected/02868e73-1352-47bd-8fb4-8697bb9e2587-kube-api-access-7k225\") on node \"crc\" DevicePath \"\"" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.606864 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/02868e73-1352-47bd-8fb4-8697bb9e2587-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.979754 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" event={"ID":"02868e73-1352-47bd-8fb4-8697bb9e2587","Type":"ContainerDied","Data":"54306a8cbe063b100fd86d1a7dfa546e70d8abdbb94657ab3ff1ae2edbd829f1"} Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.980054 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54306a8cbe063b100fd86d1a7dfa546e70d8abdbb94657ab3ff1ae2edbd829f1" Nov 27 07:22:33 crc kubenswrapper[4706]: I1127 07:22:33.979831 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.619962 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf"] Nov 27 07:22:41 crc kubenswrapper[4706]: E1127 07:22:41.620798 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerName="extract" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.620815 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerName="extract" Nov 27 07:22:41 crc kubenswrapper[4706]: E1127 07:22:41.620826 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerName="pull" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.620833 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerName="pull" Nov 27 07:22:41 crc kubenswrapper[4706]: E1127 07:22:41.620861 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerName="util" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.620869 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerName="util" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.621034 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" containerName="extract" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.621530 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.623695 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-sb7xd" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.623719 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.631415 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-apiservice-cert\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.631456 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-webhook-cert\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.631625 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv2lr\" (UniqueName: \"kubernetes.io/projected/dacd565c-fccf-46b1-a3c5-4f42366f29c6-kube-api-access-wv2lr\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.631631 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf"] Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.733168 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv2lr\" (UniqueName: \"kubernetes.io/projected/dacd565c-fccf-46b1-a3c5-4f42366f29c6-kube-api-access-wv2lr\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.733399 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-apiservice-cert\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.733451 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-webhook-cert\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.742560 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-webhook-cert\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.743015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-apiservice-cert\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.766016 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv2lr\" (UniqueName: \"kubernetes.io/projected/dacd565c-fccf-46b1-a3c5-4f42366f29c6-kube-api-access-wv2lr\") pod \"keystone-operator-controller-manager-55b75c8848-4vjhf\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:41 crc kubenswrapper[4706]: I1127 07:22:41.942496 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:42 crc kubenswrapper[4706]: I1127 07:22:42.376994 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf"] Nov 27 07:22:42 crc kubenswrapper[4706]: W1127 07:22:42.381010 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddacd565c_fccf_46b1_a3c5_4f42366f29c6.slice/crio-42e6b9a9c333256fca96fbb9a1e9a9e7d9209cd8b9797795d985512bef88073e WatchSource:0}: Error finding container 42e6b9a9c333256fca96fbb9a1e9a9e7d9209cd8b9797795d985512bef88073e: Status 404 returned error can't find the container with id 42e6b9a9c333256fca96fbb9a1e9a9e7d9209cd8b9797795d985512bef88073e Nov 27 07:22:43 crc kubenswrapper[4706]: I1127 07:22:43.054327 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" event={"ID":"dacd565c-fccf-46b1-a3c5-4f42366f29c6","Type":"ContainerStarted","Data":"42e6b9a9c333256fca96fbb9a1e9a9e7d9209cd8b9797795d985512bef88073e"} Nov 27 07:22:46 crc kubenswrapper[4706]: I1127 07:22:46.072870 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" event={"ID":"dacd565c-fccf-46b1-a3c5-4f42366f29c6","Type":"ContainerStarted","Data":"f8ca8527391e1b47007e002d069ed9ed7cc7fa669ee15bb4bcbc67af39064d05"} Nov 27 07:22:46 crc kubenswrapper[4706]: I1127 07:22:46.073209 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:46 crc kubenswrapper[4706]: I1127 07:22:46.095416 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" podStartSLOduration=1.735687682 podStartE2EDuration="5.095399527s" podCreationTimestamp="2025-11-27 07:22:41 +0000 UTC" firstStartedPulling="2025-11-27 07:22:42.38340332 +0000 UTC m=+846.272994130" lastFinishedPulling="2025-11-27 07:22:45.743115165 +0000 UTC m=+849.632705975" observedRunningTime="2025-11-27 07:22:46.093395072 +0000 UTC m=+849.982985882" watchObservedRunningTime="2025-11-27 07:22:46.095399527 +0000 UTC m=+849.984990337" Nov 27 07:22:51 crc kubenswrapper[4706]: I1127 07:22:51.946803 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:22:53 crc kubenswrapper[4706]: I1127 07:22:53.127257 4706 generic.go:334] "Generic (PLEG): container finished" podID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerID="a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f" exitCode=0 Nov 27 07:22:53 crc kubenswrapper[4706]: I1127 07:22:53.127521 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"53a204c4-9b3c-4d83-a87f-bf48abe46f77","Type":"ContainerDied","Data":"a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f"} Nov 27 07:22:54 crc kubenswrapper[4706]: I1127 07:22:54.135191 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"53a204c4-9b3c-4d83-a87f-bf48abe46f77","Type":"ContainerStarted","Data":"cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2"} Nov 27 07:22:54 crc kubenswrapper[4706]: I1127 07:22:54.135729 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.517585 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=40.362345284 podStartE2EDuration="45.517567446s" podCreationTimestamp="2025-11-27 07:22:12 +0000 UTC" firstStartedPulling="2025-11-27 07:22:14.691085825 +0000 UTC m=+818.580676665" lastFinishedPulling="2025-11-27 07:22:19.846308007 +0000 UTC m=+823.735898827" observedRunningTime="2025-11-27 07:22:54.159692913 +0000 UTC m=+858.049283723" watchObservedRunningTime="2025-11-27 07:22:57.517567446 +0000 UTC m=+861.407158256" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.518322 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs"] Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.519030 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.520810 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.526827 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs"] Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.608869 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-rwgn5"] Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.609769 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.620444 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-rwgn5"] Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.673861 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-operator-scripts\") pod \"keystone-01e7-account-create-update-f6vfs\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.673905 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zprvr\" (UniqueName: \"kubernetes.io/projected/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-kube-api-access-zprvr\") pod \"keystone-01e7-account-create-update-f6vfs\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.775473 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-operator-scripts\") pod \"keystone-db-create-rwgn5\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.775542 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-operator-scripts\") pod \"keystone-01e7-account-create-update-f6vfs\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.775570 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zprvr\" (UniqueName: \"kubernetes.io/projected/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-kube-api-access-zprvr\") pod \"keystone-01e7-account-create-update-f6vfs\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.775620 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p8q2\" (UniqueName: \"kubernetes.io/projected/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-kube-api-access-8p8q2\") pod \"keystone-db-create-rwgn5\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.776358 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-operator-scripts\") pod \"keystone-01e7-account-create-update-f6vfs\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.804608 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zprvr\" (UniqueName: \"kubernetes.io/projected/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-kube-api-access-zprvr\") pod \"keystone-01e7-account-create-update-f6vfs\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.835137 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.877076 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p8q2\" (UniqueName: \"kubernetes.io/projected/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-kube-api-access-8p8q2\") pod \"keystone-db-create-rwgn5\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.878173 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-operator-scripts\") pod \"keystone-db-create-rwgn5\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.879981 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-operator-scripts\") pod \"keystone-db-create-rwgn5\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.892381 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p8q2\" (UniqueName: \"kubernetes.io/projected/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-kube-api-access-8p8q2\") pod \"keystone-db-create-rwgn5\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:57 crc kubenswrapper[4706]: I1127 07:22:57.924211 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:22:58 crc kubenswrapper[4706]: I1127 07:22:58.092727 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs"] Nov 27 07:22:58 crc kubenswrapper[4706]: W1127 07:22:58.094917 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod680b9fbb_b235_4a8e_90d3_efbebbc3ba8f.slice/crio-f160257132ffe65f00835b447f210bdf9274d07d6c9097631c070028365a326c WatchSource:0}: Error finding container f160257132ffe65f00835b447f210bdf9274d07d6c9097631c070028365a326c: Status 404 returned error can't find the container with id f160257132ffe65f00835b447f210bdf9274d07d6c9097631c070028365a326c Nov 27 07:22:58 crc kubenswrapper[4706]: I1127 07:22:58.163745 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" event={"ID":"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f","Type":"ContainerStarted","Data":"f160257132ffe65f00835b447f210bdf9274d07d6c9097631c070028365a326c"} Nov 27 07:22:58 crc kubenswrapper[4706]: I1127 07:22:58.392658 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-rwgn5"] Nov 27 07:22:59 crc kubenswrapper[4706]: I1127 07:22:59.178091 4706 generic.go:334] "Generic (PLEG): container finished" podID="b613bb93-3d09-4cbb-9b7b-f0c9198f2412" containerID="92288fb3d75479196d597f61f708b5af55c44b0e13d806bd33700507b20cf72f" exitCode=0 Nov 27 07:22:59 crc kubenswrapper[4706]: I1127 07:22:59.178270 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-rwgn5" event={"ID":"b613bb93-3d09-4cbb-9b7b-f0c9198f2412","Type":"ContainerDied","Data":"92288fb3d75479196d597f61f708b5af55c44b0e13d806bd33700507b20cf72f"} Nov 27 07:22:59 crc kubenswrapper[4706]: I1127 07:22:59.178455 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-rwgn5" event={"ID":"b613bb93-3d09-4cbb-9b7b-f0c9198f2412","Type":"ContainerStarted","Data":"c1c77807f640aa26aef7cacd2a3e7155a7815c2aa5eae47137af4ff5112fe502"} Nov 27 07:22:59 crc kubenswrapper[4706]: I1127 07:22:59.183042 4706 generic.go:334] "Generic (PLEG): container finished" podID="680b9fbb-b235-4a8e-90d3-efbebbc3ba8f" containerID="2978ff29f5f5f9fc17133dae2cad42b5d07af45f577880c3a292eb34d02a54c3" exitCode=0 Nov 27 07:22:59 crc kubenswrapper[4706]: I1127 07:22:59.183086 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" event={"ID":"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f","Type":"ContainerDied","Data":"2978ff29f5f5f9fc17133dae2cad42b5d07af45f577880c3a292eb34d02a54c3"} Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.545481 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.550393 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.668926 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-t9n5s"] Nov 27 07:23:00 crc kubenswrapper[4706]: E1127 07:23:00.669871 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b613bb93-3d09-4cbb-9b7b-f0c9198f2412" containerName="mariadb-database-create" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.669905 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b613bb93-3d09-4cbb-9b7b-f0c9198f2412" containerName="mariadb-database-create" Nov 27 07:23:00 crc kubenswrapper[4706]: E1127 07:23:00.669944 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680b9fbb-b235-4a8e-90d3-efbebbc3ba8f" containerName="mariadb-account-create-update" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.669960 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="680b9fbb-b235-4a8e-90d3-efbebbc3ba8f" containerName="mariadb-account-create-update" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.671416 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b613bb93-3d09-4cbb-9b7b-f0c9198f2412" containerName="mariadb-database-create" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.671479 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="680b9fbb-b235-4a8e-90d3-efbebbc3ba8f" containerName="mariadb-account-create-update" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.672662 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-t9n5s" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.679680 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-nhw8z" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.686121 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-t9n5s"] Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.726198 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-operator-scripts\") pod \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.726312 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-operator-scripts\") pod \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.726376 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zprvr\" (UniqueName: \"kubernetes.io/projected/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-kube-api-access-zprvr\") pod \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\" (UID: \"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f\") " Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.726467 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p8q2\" (UniqueName: \"kubernetes.io/projected/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-kube-api-access-8p8q2\") pod \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\" (UID: \"b613bb93-3d09-4cbb-9b7b-f0c9198f2412\") " Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.727250 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b613bb93-3d09-4cbb-9b7b-f0c9198f2412" (UID: "b613bb93-3d09-4cbb-9b7b-f0c9198f2412"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.727363 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "680b9fbb-b235-4a8e-90d3-efbebbc3ba8f" (UID: "680b9fbb-b235-4a8e-90d3-efbebbc3ba8f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.731824 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-kube-api-access-zprvr" (OuterVolumeSpecName: "kube-api-access-zprvr") pod "680b9fbb-b235-4a8e-90d3-efbebbc3ba8f" (UID: "680b9fbb-b235-4a8e-90d3-efbebbc3ba8f"). InnerVolumeSpecName "kube-api-access-zprvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.732546 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-kube-api-access-8p8q2" (OuterVolumeSpecName: "kube-api-access-8p8q2") pod "b613bb93-3d09-4cbb-9b7b-f0c9198f2412" (UID: "b613bb93-3d09-4cbb-9b7b-f0c9198f2412"). InnerVolumeSpecName "kube-api-access-8p8q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.828712 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgkbd\" (UniqueName: \"kubernetes.io/projected/51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a-kube-api-access-hgkbd\") pod \"horizon-operator-index-t9n5s\" (UID: \"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a\") " pod="openstack-operators/horizon-operator-index-t9n5s" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.828812 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zprvr\" (UniqueName: \"kubernetes.io/projected/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-kube-api-access-zprvr\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.828827 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p8q2\" (UniqueName: \"kubernetes.io/projected/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-kube-api-access-8p8q2\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.828838 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.828846 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b613bb93-3d09-4cbb-9b7b-f0c9198f2412-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.930503 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgkbd\" (UniqueName: \"kubernetes.io/projected/51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a-kube-api-access-hgkbd\") pod \"horizon-operator-index-t9n5s\" (UID: \"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a\") " pod="openstack-operators/horizon-operator-index-t9n5s" Nov 27 07:23:00 crc kubenswrapper[4706]: I1127 07:23:00.945536 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgkbd\" (UniqueName: \"kubernetes.io/projected/51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a-kube-api-access-hgkbd\") pod \"horizon-operator-index-t9n5s\" (UID: \"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a\") " pod="openstack-operators/horizon-operator-index-t9n5s" Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.001387 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-t9n5s" Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.197780 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-rwgn5" event={"ID":"b613bb93-3d09-4cbb-9b7b-f0c9198f2412","Type":"ContainerDied","Data":"c1c77807f640aa26aef7cacd2a3e7155a7815c2aa5eae47137af4ff5112fe502"} Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.198034 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-rwgn5" Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.198043 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1c77807f640aa26aef7cacd2a3e7155a7815c2aa5eae47137af4ff5112fe502" Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.243528 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" event={"ID":"680b9fbb-b235-4a8e-90d3-efbebbc3ba8f","Type":"ContainerDied","Data":"f160257132ffe65f00835b447f210bdf9274d07d6c9097631c070028365a326c"} Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.243563 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f160257132ffe65f00835b447f210bdf9274d07d6c9097631c070028365a326c" Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.243627 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs" Nov 27 07:23:01 crc kubenswrapper[4706]: I1127 07:23:01.398735 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-t9n5s"] Nov 27 07:23:02 crc kubenswrapper[4706]: I1127 07:23:02.254401 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-t9n5s" event={"ID":"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a","Type":"ContainerStarted","Data":"dee5e86d3b25d9d9aa5d06dc4d0110d02e9107484a871085b1bc4d6ee8cfd5f2"} Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.062401 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-wp5gg"] Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.063544 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.065705 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-lcrdn" Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.070047 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r4md\" (UniqueName: \"kubernetes.io/projected/19bb9c9d-0e2c-443e-89c5-97987923c1f7-kube-api-access-5r4md\") pod \"swift-operator-index-wp5gg\" (UID: \"19bb9c9d-0e2c-443e-89c5-97987923c1f7\") " pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.114643 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-wp5gg"] Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.170795 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r4md\" (UniqueName: \"kubernetes.io/projected/19bb9c9d-0e2c-443e-89c5-97987923c1f7-kube-api-access-5r4md\") pod \"swift-operator-index-wp5gg\" (UID: \"19bb9c9d-0e2c-443e-89c5-97987923c1f7\") " pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.191889 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r4md\" (UniqueName: \"kubernetes.io/projected/19bb9c9d-0e2c-443e-89c5-97987923c1f7-kube-api-access-5r4md\") pod \"swift-operator-index-wp5gg\" (UID: \"19bb9c9d-0e2c-443e-89c5-97987923c1f7\") " pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.261679 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-t9n5s" event={"ID":"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a","Type":"ContainerStarted","Data":"6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20"} Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.275890 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-t9n5s" podStartSLOduration=2.398109027 podStartE2EDuration="3.275869146s" podCreationTimestamp="2025-11-27 07:23:00 +0000 UTC" firstStartedPulling="2025-11-27 07:23:01.40577807 +0000 UTC m=+865.295368900" lastFinishedPulling="2025-11-27 07:23:02.283538199 +0000 UTC m=+866.173129019" observedRunningTime="2025-11-27 07:23:03.274051546 +0000 UTC m=+867.163642366" watchObservedRunningTime="2025-11-27 07:23:03.275869146 +0000 UTC m=+867.165459956" Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.393647 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:03 crc kubenswrapper[4706]: I1127 07:23:03.888084 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-wp5gg"] Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.234361 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.284202 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-wp5gg" event={"ID":"19bb9c9d-0e2c-443e-89c5-97987923c1f7","Type":"ContainerStarted","Data":"7d421fdcdf5bb347cc86539ed779ba32045f1e95d0b9548f56bf709a36f4476e"} Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.786560 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-fqrtx"] Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.787753 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.792134 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-z66nv" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.792463 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.792752 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.792906 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.794232 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-fqrtx"] Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.894068 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-config-data\") pod \"keystone-db-sync-fqrtx\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.894333 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh5x8\" (UniqueName: \"kubernetes.io/projected/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-kube-api-access-fh5x8\") pod \"keystone-db-sync-fqrtx\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.995354 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-config-data\") pod \"keystone-db-sync-fqrtx\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:04 crc kubenswrapper[4706]: I1127 07:23:04.995449 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh5x8\" (UniqueName: \"kubernetes.io/projected/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-kube-api-access-fh5x8\") pod \"keystone-db-sync-fqrtx\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:05 crc kubenswrapper[4706]: I1127 07:23:05.002046 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-config-data\") pod \"keystone-db-sync-fqrtx\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:05 crc kubenswrapper[4706]: I1127 07:23:05.010685 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh5x8\" (UniqueName: \"kubernetes.io/projected/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-kube-api-access-fh5x8\") pod \"keystone-db-sync-fqrtx\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:05 crc kubenswrapper[4706]: I1127 07:23:05.116153 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:05 crc kubenswrapper[4706]: I1127 07:23:05.295511 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-wp5gg" event={"ID":"19bb9c9d-0e2c-443e-89c5-97987923c1f7","Type":"ContainerStarted","Data":"3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626"} Nov 27 07:23:05 crc kubenswrapper[4706]: I1127 07:23:05.313203 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-wp5gg" podStartSLOduration=1.337203606 podStartE2EDuration="2.313190163s" podCreationTimestamp="2025-11-27 07:23:03 +0000 UTC" firstStartedPulling="2025-11-27 07:23:03.900474056 +0000 UTC m=+867.790064866" lastFinishedPulling="2025-11-27 07:23:04.876460613 +0000 UTC m=+868.766051423" observedRunningTime="2025-11-27 07:23:05.311454475 +0000 UTC m=+869.201045285" watchObservedRunningTime="2025-11-27 07:23:05.313190163 +0000 UTC m=+869.202780973" Nov 27 07:23:05 crc kubenswrapper[4706]: I1127 07:23:05.516244 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-fqrtx"] Nov 27 07:23:05 crc kubenswrapper[4706]: W1127 07:23:05.525196 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0685d9b5_6dde_48e8_ac3d_4d247c5ecf06.slice/crio-abb8a03f93d7c9e5e11d4dfd5d33088365a517b332d1eaac36c7f32a16be1ff5 WatchSource:0}: Error finding container abb8a03f93d7c9e5e11d4dfd5d33088365a517b332d1eaac36c7f32a16be1ff5: Status 404 returned error can't find the container with id abb8a03f93d7c9e5e11d4dfd5d33088365a517b332d1eaac36c7f32a16be1ff5 Nov 27 07:23:06 crc kubenswrapper[4706]: I1127 07:23:06.305336 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" event={"ID":"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06","Type":"ContainerStarted","Data":"abb8a03f93d7c9e5e11d4dfd5d33088365a517b332d1eaac36c7f32a16be1ff5"} Nov 27 07:23:06 crc kubenswrapper[4706]: I1127 07:23:06.866976 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-t9n5s"] Nov 27 07:23:06 crc kubenswrapper[4706]: I1127 07:23:06.867510 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/horizon-operator-index-t9n5s" podUID="51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a" containerName="registry-server" containerID="cri-o://6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20" gracePeriod=2 Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.259529 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-t9n5s" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.315961 4706 generic.go:334] "Generic (PLEG): container finished" podID="51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a" containerID="6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20" exitCode=0 Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.316006 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-t9n5s" event={"ID":"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a","Type":"ContainerDied","Data":"6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20"} Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.316032 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-t9n5s" event={"ID":"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a","Type":"ContainerDied","Data":"dee5e86d3b25d9d9aa5d06dc4d0110d02e9107484a871085b1bc4d6ee8cfd5f2"} Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.316048 4706 scope.go:117] "RemoveContainer" containerID="6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.316209 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-t9n5s" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.328332 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgkbd\" (UniqueName: \"kubernetes.io/projected/51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a-kube-api-access-hgkbd\") pod \"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a\" (UID: \"51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a\") " Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.334512 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a-kube-api-access-hgkbd" (OuterVolumeSpecName: "kube-api-access-hgkbd") pod "51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a" (UID: "51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a"). InnerVolumeSpecName "kube-api-access-hgkbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.338252 4706 scope.go:117] "RemoveContainer" containerID="6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20" Nov 27 07:23:07 crc kubenswrapper[4706]: E1127 07:23:07.338725 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20\": container with ID starting with 6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20 not found: ID does not exist" containerID="6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.338764 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20"} err="failed to get container status \"6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20\": rpc error: code = NotFound desc = could not find container \"6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20\": container with ID starting with 6520abef38f115948f595b5f52f8b6079f187d2359ccf9f083523178b6b5cb20 not found: ID does not exist" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.429420 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgkbd\" (UniqueName: \"kubernetes.io/projected/51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a-kube-api-access-hgkbd\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.462128 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-jnvft"] Nov 27 07:23:07 crc kubenswrapper[4706]: E1127 07:23:07.464307 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a" containerName="registry-server" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.464331 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a" containerName="registry-server" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.464600 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a" containerName="registry-server" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.465658 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.470752 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-jnvft"] Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.531410 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn89c\" (UniqueName: \"kubernetes.io/projected/7e5fd17c-d97a-4826-bd68-cd8480f1363d-kube-api-access-pn89c\") pod \"horizon-operator-index-jnvft\" (UID: \"7e5fd17c-d97a-4826-bd68-cd8480f1363d\") " pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.632907 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn89c\" (UniqueName: \"kubernetes.io/projected/7e5fd17c-d97a-4826-bd68-cd8480f1363d-kube-api-access-pn89c\") pod \"horizon-operator-index-jnvft\" (UID: \"7e5fd17c-d97a-4826-bd68-cd8480f1363d\") " pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.649552 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn89c\" (UniqueName: \"kubernetes.io/projected/7e5fd17c-d97a-4826-bd68-cd8480f1363d-kube-api-access-pn89c\") pod \"horizon-operator-index-jnvft\" (UID: \"7e5fd17c-d97a-4826-bd68-cd8480f1363d\") " pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.661844 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-t9n5s"] Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.667176 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/horizon-operator-index-t9n5s"] Nov 27 07:23:07 crc kubenswrapper[4706]: I1127 07:23:07.786892 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:08 crc kubenswrapper[4706]: I1127 07:23:08.157207 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-jnvft"] Nov 27 07:23:08 crc kubenswrapper[4706]: I1127 07:23:08.790984 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a" path="/var/lib/kubelet/pods/51e2af2f-ba5a-4f4c-aa3a-aeb9c78ddd9a/volumes" Nov 27 07:23:10 crc kubenswrapper[4706]: W1127 07:23:10.751591 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e5fd17c_d97a_4826_bd68_cd8480f1363d.slice/crio-43a5058326dd2735c93856272fb6b12922fab79dad477e65da34ae92bb1499cd WatchSource:0}: Error finding container 43a5058326dd2735c93856272fb6b12922fab79dad477e65da34ae92bb1499cd: Status 404 returned error can't find the container with id 43a5058326dd2735c93856272fb6b12922fab79dad477e65da34ae92bb1499cd Nov 27 07:23:11 crc kubenswrapper[4706]: I1127 07:23:11.347276 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-jnvft" event={"ID":"7e5fd17c-d97a-4826-bd68-cd8480f1363d","Type":"ContainerStarted","Data":"43a5058326dd2735c93856272fb6b12922fab79dad477e65da34ae92bb1499cd"} Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.394208 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.394288 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.425790 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.668022 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dk9jh"] Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.669510 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.688790 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk9jh"] Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.824211 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c76c\" (UniqueName: \"kubernetes.io/projected/150d6d02-355e-4966-94be-43e8058068de-kube-api-access-9c76c\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.824313 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-utilities\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.824345 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-catalog-content\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.925997 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c76c\" (UniqueName: \"kubernetes.io/projected/150d6d02-355e-4966-94be-43e8058068de-kube-api-access-9c76c\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.926076 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-utilities\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.926126 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-catalog-content\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.926625 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-utilities\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.926667 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-catalog-content\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.947868 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c76c\" (UniqueName: \"kubernetes.io/projected/150d6d02-355e-4966-94be-43e8058068de-kube-api-access-9c76c\") pod \"certified-operators-dk9jh\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:13 crc kubenswrapper[4706]: I1127 07:23:13.993408 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:14 crc kubenswrapper[4706]: I1127 07:23:14.394510 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:23:15 crc kubenswrapper[4706]: I1127 07:23:15.178259 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:23:15 crc kubenswrapper[4706]: I1127 07:23:15.178616 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.038128 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk9jh"] Nov 27 07:23:16 crc kubenswrapper[4706]: W1127 07:23:16.042030 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod150d6d02_355e_4966_94be_43e8058068de.slice/crio-6a111ed695c97fad5e19b75d05534a82f7cb73af56dc275954a8938f0435dd5f WatchSource:0}: Error finding container 6a111ed695c97fad5e19b75d05534a82f7cb73af56dc275954a8938f0435dd5f: Status 404 returned error can't find the container with id 6a111ed695c97fad5e19b75d05534a82f7cb73af56dc275954a8938f0435dd5f Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.380530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-jnvft" event={"ID":"7e5fd17c-d97a-4826-bd68-cd8480f1363d","Type":"ContainerStarted","Data":"d9e204fe947b7e250a46705ebc28008fb05c7674cf8ddefcd1c14983d492a1ae"} Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.382190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" event={"ID":"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06","Type":"ContainerStarted","Data":"6fb01b105045e54c0460e23063053254e968a28290766f6c9c688dc0c5ec831b"} Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.383772 4706 generic.go:334] "Generic (PLEG): container finished" podID="150d6d02-355e-4966-94be-43e8058068de" containerID="a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703" exitCode=0 Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.383808 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk9jh" event={"ID":"150d6d02-355e-4966-94be-43e8058068de","Type":"ContainerDied","Data":"a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703"} Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.383826 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk9jh" event={"ID":"150d6d02-355e-4966-94be-43e8058068de","Type":"ContainerStarted","Data":"6a111ed695c97fad5e19b75d05534a82f7cb73af56dc275954a8938f0435dd5f"} Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.396118 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-jnvft" podStartSLOduration=6.7939223890000005 podStartE2EDuration="9.396097948s" podCreationTimestamp="2025-11-27 07:23:07 +0000 UTC" firstStartedPulling="2025-11-27 07:23:13.022502161 +0000 UTC m=+876.912092971" lastFinishedPulling="2025-11-27 07:23:15.6246777 +0000 UTC m=+879.514268530" observedRunningTime="2025-11-27 07:23:16.393493676 +0000 UTC m=+880.283084496" watchObservedRunningTime="2025-11-27 07:23:16.396097948 +0000 UTC m=+880.285688758" Nov 27 07:23:16 crc kubenswrapper[4706]: I1127 07:23:16.416071 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" podStartSLOduration=2.285017196 podStartE2EDuration="12.416045658s" podCreationTimestamp="2025-11-27 07:23:04 +0000 UTC" firstStartedPulling="2025-11-27 07:23:05.527295606 +0000 UTC m=+869.416886416" lastFinishedPulling="2025-11-27 07:23:15.658324068 +0000 UTC m=+879.547914878" observedRunningTime="2025-11-27 07:23:16.411927764 +0000 UTC m=+880.301518594" watchObservedRunningTime="2025-11-27 07:23:16.416045658 +0000 UTC m=+880.305636478" Nov 27 07:23:17 crc kubenswrapper[4706]: I1127 07:23:17.787034 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:17 crc kubenswrapper[4706]: I1127 07:23:17.787537 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:17 crc kubenswrapper[4706]: I1127 07:23:17.822058 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:18 crc kubenswrapper[4706]: I1127 07:23:18.395498 4706 generic.go:334] "Generic (PLEG): container finished" podID="150d6d02-355e-4966-94be-43e8058068de" containerID="0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164" exitCode=0 Nov 27 07:23:18 crc kubenswrapper[4706]: I1127 07:23:18.395573 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk9jh" event={"ID":"150d6d02-355e-4966-94be-43e8058068de","Type":"ContainerDied","Data":"0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164"} Nov 27 07:23:19 crc kubenswrapper[4706]: I1127 07:23:19.403867 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk9jh" event={"ID":"150d6d02-355e-4966-94be-43e8058068de","Type":"ContainerStarted","Data":"b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04"} Nov 27 07:23:19 crc kubenswrapper[4706]: I1127 07:23:19.407201 4706 generic.go:334] "Generic (PLEG): container finished" podID="0685d9b5-6dde-48e8-ac3d-4d247c5ecf06" containerID="6fb01b105045e54c0460e23063053254e968a28290766f6c9c688dc0c5ec831b" exitCode=0 Nov 27 07:23:19 crc kubenswrapper[4706]: I1127 07:23:19.407248 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" event={"ID":"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06","Type":"ContainerDied","Data":"6fb01b105045e54c0460e23063053254e968a28290766f6c9c688dc0c5ec831b"} Nov 27 07:23:19 crc kubenswrapper[4706]: I1127 07:23:19.427070 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dk9jh" podStartSLOduration=3.8291329960000002 podStartE2EDuration="6.427052978s" podCreationTimestamp="2025-11-27 07:23:13 +0000 UTC" firstStartedPulling="2025-11-27 07:23:16.384725364 +0000 UTC m=+880.274316174" lastFinishedPulling="2025-11-27 07:23:18.982645346 +0000 UTC m=+882.872236156" observedRunningTime="2025-11-27 07:23:19.42457485 +0000 UTC m=+883.314165660" watchObservedRunningTime="2025-11-27 07:23:19.427052978 +0000 UTC m=+883.316643798" Nov 27 07:23:20 crc kubenswrapper[4706]: I1127 07:23:20.751444 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:20 crc kubenswrapper[4706]: I1127 07:23:20.823795 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-config-data\") pod \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " Nov 27 07:23:20 crc kubenswrapper[4706]: I1127 07:23:20.824174 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh5x8\" (UniqueName: \"kubernetes.io/projected/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-kube-api-access-fh5x8\") pod \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\" (UID: \"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06\") " Nov 27 07:23:20 crc kubenswrapper[4706]: I1127 07:23:20.829403 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-kube-api-access-fh5x8" (OuterVolumeSpecName: "kube-api-access-fh5x8") pod "0685d9b5-6dde-48e8-ac3d-4d247c5ecf06" (UID: "0685d9b5-6dde-48e8-ac3d-4d247c5ecf06"). InnerVolumeSpecName "kube-api-access-fh5x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:20 crc kubenswrapper[4706]: I1127 07:23:20.862169 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-config-data" (OuterVolumeSpecName: "config-data") pod "0685d9b5-6dde-48e8-ac3d-4d247c5ecf06" (UID: "0685d9b5-6dde-48e8-ac3d-4d247c5ecf06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:23:20 crc kubenswrapper[4706]: I1127 07:23:20.925604 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh5x8\" (UniqueName: \"kubernetes.io/projected/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-kube-api-access-fh5x8\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:20 crc kubenswrapper[4706]: I1127 07:23:20.925637 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.423277 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" event={"ID":"0685d9b5-6dde-48e8-ac3d-4d247c5ecf06","Type":"ContainerDied","Data":"abb8a03f93d7c9e5e11d4dfd5d33088365a517b332d1eaac36c7f32a16be1ff5"} Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.423338 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-fqrtx" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.423348 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abb8a03f93d7c9e5e11d4dfd5d33088365a517b332d1eaac36c7f32a16be1ff5" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.656665 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-m4zj7"] Nov 27 07:23:21 crc kubenswrapper[4706]: E1127 07:23:21.656897 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0685d9b5-6dde-48e8-ac3d-4d247c5ecf06" containerName="keystone-db-sync" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.656911 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0685d9b5-6dde-48e8-ac3d-4d247c5ecf06" containerName="keystone-db-sync" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.657033 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0685d9b5-6dde-48e8-ac3d-4d247c5ecf06" containerName="keystone-db-sync" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.657438 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.661822 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-z66nv" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.662016 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.664895 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.665183 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.665480 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.692309 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-m4zj7"] Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.736280 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-config-data\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.736398 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-scripts\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.736455 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-fernet-keys\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.736609 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzv44\" (UniqueName: \"kubernetes.io/projected/c827e889-9073-49b1-a264-bfa826083f54-kube-api-access-qzv44\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.736672 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-credential-keys\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.838161 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzv44\" (UniqueName: \"kubernetes.io/projected/c827e889-9073-49b1-a264-bfa826083f54-kube-api-access-qzv44\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.838437 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-credential-keys\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.838647 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-config-data\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.839094 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-scripts\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.839421 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-fernet-keys\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.845907 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-config-data\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.847256 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-scripts\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.848330 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-credential-keys\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.851593 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-fernet-keys\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.855121 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzv44\" (UniqueName: \"kubernetes.io/projected/c827e889-9073-49b1-a264-bfa826083f54-kube-api-access-qzv44\") pod \"keystone-bootstrap-m4zj7\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:21 crc kubenswrapper[4706]: I1127 07:23:21.982083 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:22 crc kubenswrapper[4706]: I1127 07:23:22.512181 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-m4zj7"] Nov 27 07:23:23 crc kubenswrapper[4706]: I1127 07:23:23.437615 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" event={"ID":"c827e889-9073-49b1-a264-bfa826083f54","Type":"ContainerStarted","Data":"4a52d9d607e438f583ea3517ffab5bd0bc362f7711bb9bbb1fbb1f90aeebd843"} Nov 27 07:23:23 crc kubenswrapper[4706]: I1127 07:23:23.438743 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" event={"ID":"c827e889-9073-49b1-a264-bfa826083f54","Type":"ContainerStarted","Data":"12c63ef06da993f15da3abb876758bbe4281074dd0326b8fc39b91627c3812c1"} Nov 27 07:23:23 crc kubenswrapper[4706]: I1127 07:23:23.460770 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" podStartSLOduration=2.460756281 podStartE2EDuration="2.460756281s" podCreationTimestamp="2025-11-27 07:23:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:23:23.458629224 +0000 UTC m=+887.348220044" watchObservedRunningTime="2025-11-27 07:23:23.460756281 +0000 UTC m=+887.350347091" Nov 27 07:23:23 crc kubenswrapper[4706]: I1127 07:23:23.994323 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:23 crc kubenswrapper[4706]: I1127 07:23:23.994610 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:24 crc kubenswrapper[4706]: I1127 07:23:24.038217 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:24 crc kubenswrapper[4706]: I1127 07:23:24.501749 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.457417 4706 generic.go:334] "Generic (PLEG): container finished" podID="c827e889-9073-49b1-a264-bfa826083f54" containerID="4a52d9d607e438f583ea3517ffab5bd0bc362f7711bb9bbb1fbb1f90aeebd843" exitCode=0 Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.457490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" event={"ID":"c827e889-9073-49b1-a264-bfa826083f54","Type":"ContainerDied","Data":"4a52d9d607e438f583ea3517ffab5bd0bc362f7711bb9bbb1fbb1f90aeebd843"} Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.868276 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cxkkm"] Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.870103 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.886784 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxkkm"] Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.995008 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-catalog-content\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.995170 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-utilities\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:25 crc kubenswrapper[4706]: I1127 07:23:25.995206 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmhq8\" (UniqueName: \"kubernetes.io/projected/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-kube-api-access-tmhq8\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.096106 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-catalog-content\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.096603 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-catalog-content\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.096977 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-utilities\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.097103 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmhq8\" (UniqueName: \"kubernetes.io/projected/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-kube-api-access-tmhq8\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.097385 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-utilities\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.116934 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmhq8\" (UniqueName: \"kubernetes.io/projected/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-kube-api-access-tmhq8\") pod \"redhat-marketplace-cxkkm\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.244065 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.657210 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxkkm"] Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.822801 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.906077 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzv44\" (UniqueName: \"kubernetes.io/projected/c827e889-9073-49b1-a264-bfa826083f54-kube-api-access-qzv44\") pod \"c827e889-9073-49b1-a264-bfa826083f54\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.906139 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-fernet-keys\") pod \"c827e889-9073-49b1-a264-bfa826083f54\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.906188 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-config-data\") pod \"c827e889-9073-49b1-a264-bfa826083f54\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.906304 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-credential-keys\") pod \"c827e889-9073-49b1-a264-bfa826083f54\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.906334 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-scripts\") pod \"c827e889-9073-49b1-a264-bfa826083f54\" (UID: \"c827e889-9073-49b1-a264-bfa826083f54\") " Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.910756 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-scripts" (OuterVolumeSpecName: "scripts") pod "c827e889-9073-49b1-a264-bfa826083f54" (UID: "c827e889-9073-49b1-a264-bfa826083f54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.910892 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c827e889-9073-49b1-a264-bfa826083f54" (UID: "c827e889-9073-49b1-a264-bfa826083f54"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.911127 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c827e889-9073-49b1-a264-bfa826083f54-kube-api-access-qzv44" (OuterVolumeSpecName: "kube-api-access-qzv44") pod "c827e889-9073-49b1-a264-bfa826083f54" (UID: "c827e889-9073-49b1-a264-bfa826083f54"). InnerVolumeSpecName "kube-api-access-qzv44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.911823 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c827e889-9073-49b1-a264-bfa826083f54" (UID: "c827e889-9073-49b1-a264-bfa826083f54"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:23:26 crc kubenswrapper[4706]: I1127 07:23:26.925868 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-config-data" (OuterVolumeSpecName: "config-data") pod "c827e889-9073-49b1-a264-bfa826083f54" (UID: "c827e889-9073-49b1-a264-bfa826083f54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.008438 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzv44\" (UniqueName: \"kubernetes.io/projected/c827e889-9073-49b1-a264-bfa826083f54-kube-api-access-qzv44\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.008481 4706 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.008541 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.008583 4706 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.008596 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c827e889-9073-49b1-a264-bfa826083f54-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.265191 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dk9jh"] Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.473150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" event={"ID":"c827e889-9073-49b1-a264-bfa826083f54","Type":"ContainerDied","Data":"12c63ef06da993f15da3abb876758bbe4281074dd0326b8fc39b91627c3812c1"} Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.473183 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-m4zj7" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.473199 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12c63ef06da993f15da3abb876758bbe4281074dd0326b8fc39b91627c3812c1" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.474705 4706 generic.go:334] "Generic (PLEG): container finished" podID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerID="d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9" exitCode=0 Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.474791 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxkkm" event={"ID":"33052f7e-58dc-472e-8ffd-2e6fd9c16c27","Type":"ContainerDied","Data":"d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9"} Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.474822 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxkkm" event={"ID":"33052f7e-58dc-472e-8ffd-2e6fd9c16c27","Type":"ContainerStarted","Data":"be8d58a05bad5371a9fe265d903195f7d59d9cfe0f130de298f1d2f317e5e031"} Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.474955 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dk9jh" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="registry-server" containerID="cri-o://b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04" gracePeriod=2 Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.646951 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-59949565db-f2499"] Nov 27 07:23:27 crc kubenswrapper[4706]: E1127 07:23:27.647186 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c827e889-9073-49b1-a264-bfa826083f54" containerName="keystone-bootstrap" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.647199 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c827e889-9073-49b1-a264-bfa826083f54" containerName="keystone-bootstrap" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.647329 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c827e889-9073-49b1-a264-bfa826083f54" containerName="keystone-bootstrap" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.647762 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.650756 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.651012 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.651198 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.655542 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-z66nv" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.663727 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-59949565db-f2499"] Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.717724 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-scripts\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.717781 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-config-data\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.717824 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-fernet-keys\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.717844 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp4dt\" (UniqueName: \"kubernetes.io/projected/b67a1baa-0133-4b6d-b490-20ff47b1f90f-kube-api-access-tp4dt\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.717874 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-credential-keys\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.815079 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-jnvft" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.819135 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-credential-keys\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.819231 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-scripts\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.819261 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-config-data\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.819292 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp4dt\" (UniqueName: \"kubernetes.io/projected/b67a1baa-0133-4b6d-b490-20ff47b1f90f-kube-api-access-tp4dt\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.819307 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-fernet-keys\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.824100 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-scripts\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.824382 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-fernet-keys\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.824452 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-credential-keys\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.827326 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-config-data\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.852791 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp4dt\" (UniqueName: \"kubernetes.io/projected/b67a1baa-0133-4b6d-b490-20ff47b1f90f-kube-api-access-tp4dt\") pod \"keystone-59949565db-f2499\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.883482 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:27 crc kubenswrapper[4706]: I1127 07:23:27.974272 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.021664 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-catalog-content\") pod \"150d6d02-355e-4966-94be-43e8058068de\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.021766 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-utilities\") pod \"150d6d02-355e-4966-94be-43e8058068de\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.021902 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c76c\" (UniqueName: \"kubernetes.io/projected/150d6d02-355e-4966-94be-43e8058068de-kube-api-access-9c76c\") pod \"150d6d02-355e-4966-94be-43e8058068de\" (UID: \"150d6d02-355e-4966-94be-43e8058068de\") " Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.026316 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/150d6d02-355e-4966-94be-43e8058068de-kube-api-access-9c76c" (OuterVolumeSpecName: "kube-api-access-9c76c") pod "150d6d02-355e-4966-94be-43e8058068de" (UID: "150d6d02-355e-4966-94be-43e8058068de"). InnerVolumeSpecName "kube-api-access-9c76c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.026430 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-utilities" (OuterVolumeSpecName: "utilities") pod "150d6d02-355e-4966-94be-43e8058068de" (UID: "150d6d02-355e-4966-94be-43e8058068de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.123507 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.123833 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c76c\" (UniqueName: \"kubernetes.io/projected/150d6d02-355e-4966-94be-43e8058068de-kube-api-access-9c76c\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.150852 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "150d6d02-355e-4966-94be-43e8058068de" (UID: "150d6d02-355e-4966-94be-43e8058068de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.184486 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-59949565db-f2499"] Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.225108 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150d6d02-355e-4966-94be-43e8058068de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.483101 4706 generic.go:334] "Generic (PLEG): container finished" podID="150d6d02-355e-4966-94be-43e8058068de" containerID="b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04" exitCode=0 Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.483147 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk9jh" event={"ID":"150d6d02-355e-4966-94be-43e8058068de","Type":"ContainerDied","Data":"b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04"} Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.483549 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk9jh" event={"ID":"150d6d02-355e-4966-94be-43e8058068de","Type":"ContainerDied","Data":"6a111ed695c97fad5e19b75d05534a82f7cb73af56dc275954a8938f0435dd5f"} Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.483166 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk9jh" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.483570 4706 scope.go:117] "RemoveContainer" containerID="b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.485570 4706 generic.go:334] "Generic (PLEG): container finished" podID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerID="8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6" exitCode=0 Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.485607 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxkkm" event={"ID":"33052f7e-58dc-472e-8ffd-2e6fd9c16c27","Type":"ContainerDied","Data":"8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6"} Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.487310 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-59949565db-f2499" event={"ID":"b67a1baa-0133-4b6d-b490-20ff47b1f90f","Type":"ContainerStarted","Data":"5c0d5aeb797927517a277670205f5bba69b69629a355e0c670e0024c294bc4bd"} Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.487345 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-59949565db-f2499" event={"ID":"b67a1baa-0133-4b6d-b490-20ff47b1f90f","Type":"ContainerStarted","Data":"c2347c5e3516f02865f03f40a4415dfe61cc7eb79e6ff9067ec74c932a31bb33"} Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.487442 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.503003 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-59949565db-f2499" podStartSLOduration=1.502986231 podStartE2EDuration="1.502986231s" podCreationTimestamp="2025-11-27 07:23:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:23:28.501887397 +0000 UTC m=+892.391478217" watchObservedRunningTime="2025-11-27 07:23:28.502986231 +0000 UTC m=+892.392577041" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.519952 4706 scope.go:117] "RemoveContainer" containerID="0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.537493 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dk9jh"] Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.541813 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dk9jh"] Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.557397 4706 scope.go:117] "RemoveContainer" containerID="a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.570324 4706 scope.go:117] "RemoveContainer" containerID="b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04" Nov 27 07:23:28 crc kubenswrapper[4706]: E1127 07:23:28.570649 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04\": container with ID starting with b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04 not found: ID does not exist" containerID="b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.570681 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04"} err="failed to get container status \"b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04\": rpc error: code = NotFound desc = could not find container \"b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04\": container with ID starting with b9d3e523d29159fe36e826fb0246651f1a4e954e6442e75e9b70332a3b5bbf04 not found: ID does not exist" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.570704 4706 scope.go:117] "RemoveContainer" containerID="0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164" Nov 27 07:23:28 crc kubenswrapper[4706]: E1127 07:23:28.571054 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164\": container with ID starting with 0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164 not found: ID does not exist" containerID="0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.571073 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164"} err="failed to get container status \"0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164\": rpc error: code = NotFound desc = could not find container \"0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164\": container with ID starting with 0210bf93787d9ea99af649643695b9016f88fe0106831d16484283191b2bd164 not found: ID does not exist" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.571088 4706 scope.go:117] "RemoveContainer" containerID="a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703" Nov 27 07:23:28 crc kubenswrapper[4706]: E1127 07:23:28.571695 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703\": container with ID starting with a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703 not found: ID does not exist" containerID="a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.571730 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703"} err="failed to get container status \"a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703\": rpc error: code = NotFound desc = could not find container \"a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703\": container with ID starting with a9cbb2ecf721ac080c42f2555f7de1fc5a154621c9bfc3d649b7ace7c63d5703 not found: ID does not exist" Nov 27 07:23:28 crc kubenswrapper[4706]: I1127 07:23:28.786072 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="150d6d02-355e-4966-94be-43e8058068de" path="/var/lib/kubelet/pods/150d6d02-355e-4966-94be-43e8058068de/volumes" Nov 27 07:23:29 crc kubenswrapper[4706]: I1127 07:23:29.498111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxkkm" event={"ID":"33052f7e-58dc-472e-8ffd-2e6fd9c16c27","Type":"ContainerStarted","Data":"525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422"} Nov 27 07:23:29 crc kubenswrapper[4706]: I1127 07:23:29.522053 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cxkkm" podStartSLOduration=2.817739166 podStartE2EDuration="4.522025615s" podCreationTimestamp="2025-11-27 07:23:25 +0000 UTC" firstStartedPulling="2025-11-27 07:23:27.476638256 +0000 UTC m=+891.366229066" lastFinishedPulling="2025-11-27 07:23:29.180924705 +0000 UTC m=+893.070515515" observedRunningTime="2025-11-27 07:23:29.517979207 +0000 UTC m=+893.407570017" watchObservedRunningTime="2025-11-27 07:23:29.522025615 +0000 UTC m=+893.411616425" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.068938 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pblwf"] Nov 27 07:23:30 crc kubenswrapper[4706]: E1127 07:23:30.069162 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="registry-server" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.069174 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="registry-server" Nov 27 07:23:30 crc kubenswrapper[4706]: E1127 07:23:30.069194 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="extract-content" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.069200 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="extract-content" Nov 27 07:23:30 crc kubenswrapper[4706]: E1127 07:23:30.069210 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="extract-utilities" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.069234 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="extract-utilities" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.069343 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="150d6d02-355e-4966-94be-43e8058068de" containerName="registry-server" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.070159 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.090684 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pblwf"] Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.157903 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-catalog-content\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.157985 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnn4k\" (UniqueName: \"kubernetes.io/projected/49808ffe-83fb-4aa7-8eae-8debb8354286-kube-api-access-lnn4k\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.158101 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-utilities\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.259969 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-utilities\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.260058 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-catalog-content\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.260106 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnn4k\" (UniqueName: \"kubernetes.io/projected/49808ffe-83fb-4aa7-8eae-8debb8354286-kube-api-access-lnn4k\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.260524 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-utilities\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.260560 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-catalog-content\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.285653 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnn4k\" (UniqueName: \"kubernetes.io/projected/49808ffe-83fb-4aa7-8eae-8debb8354286-kube-api-access-lnn4k\") pod \"community-operators-pblwf\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.389712 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:30 crc kubenswrapper[4706]: I1127 07:23:30.825311 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pblwf"] Nov 27 07:23:31 crc kubenswrapper[4706]: I1127 07:23:31.515946 4706 generic.go:334] "Generic (PLEG): container finished" podID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerID="17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d" exitCode=0 Nov 27 07:23:31 crc kubenswrapper[4706]: I1127 07:23:31.516002 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pblwf" event={"ID":"49808ffe-83fb-4aa7-8eae-8debb8354286","Type":"ContainerDied","Data":"17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d"} Nov 27 07:23:31 crc kubenswrapper[4706]: I1127 07:23:31.516271 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pblwf" event={"ID":"49808ffe-83fb-4aa7-8eae-8debb8354286","Type":"ContainerStarted","Data":"da73a168cb03772a3c5cd9d60e161d6aaedcc4eb2b3795736ec3e04872ff389e"} Nov 27 07:23:32 crc kubenswrapper[4706]: I1127 07:23:32.524303 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pblwf" event={"ID":"49808ffe-83fb-4aa7-8eae-8debb8354286","Type":"ContainerStarted","Data":"9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc"} Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.534202 4706 generic.go:334] "Generic (PLEG): container finished" podID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerID="9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc" exitCode=0 Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.534274 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pblwf" event={"ID":"49808ffe-83fb-4aa7-8eae-8debb8354286","Type":"ContainerDied","Data":"9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc"} Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.710370 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9"] Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.711595 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.713926 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dk9r5" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.722584 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9"] Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.814106 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.814166 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.814263 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx2bq\" (UniqueName: \"kubernetes.io/projected/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-kube-api-access-tx2bq\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.915886 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.915956 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.916494 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx2bq\" (UniqueName: \"kubernetes.io/projected/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-kube-api-access-tx2bq\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.916546 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.916643 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:33 crc kubenswrapper[4706]: I1127 07:23:33.946032 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx2bq\" (UniqueName: \"kubernetes.io/projected/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-kube-api-access-tx2bq\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.041555 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.384615 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9"] Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.541922 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" event={"ID":"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee","Type":"ContainerStarted","Data":"fa2a6b2b83c7e6262e1e8885b82b53deb1a8f3687cc2e5205bd39b03c2dd89ce"} Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.542901 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" event={"ID":"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee","Type":"ContainerStarted","Data":"f33a904ac5f2541327d76425d5581734fb62c12a06cff11fda1baf6bfc6b3613"} Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.544338 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pblwf" event={"ID":"49808ffe-83fb-4aa7-8eae-8debb8354286","Type":"ContainerStarted","Data":"93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978"} Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.584463 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pblwf" podStartSLOduration=2.155150059 podStartE2EDuration="4.584444635s" podCreationTimestamp="2025-11-27 07:23:30 +0000 UTC" firstStartedPulling="2025-11-27 07:23:31.517405029 +0000 UTC m=+895.406995859" lastFinishedPulling="2025-11-27 07:23:33.946699625 +0000 UTC m=+897.836290435" observedRunningTime="2025-11-27 07:23:34.577450373 +0000 UTC m=+898.467041183" watchObservedRunningTime="2025-11-27 07:23:34.584444635 +0000 UTC m=+898.474035445" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.702234 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6"] Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.703507 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.751563 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6"] Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.833668 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5gm4\" (UniqueName: \"kubernetes.io/projected/82d9ba46-8fc0-462c-bc77-056ef3252965-kube-api-access-w5gm4\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.833716 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.833797 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.935394 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5gm4\" (UniqueName: \"kubernetes.io/projected/82d9ba46-8fc0-462c-bc77-056ef3252965-kube-api-access-w5gm4\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.935438 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.935490 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.935874 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.936329 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:34 crc kubenswrapper[4706]: I1127 07:23:34.955498 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5gm4\" (UniqueName: \"kubernetes.io/projected/82d9ba46-8fc0-462c-bc77-056ef3252965-kube-api-access-w5gm4\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:35 crc kubenswrapper[4706]: I1127 07:23:35.023580 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:35 crc kubenswrapper[4706]: I1127 07:23:35.234841 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6"] Nov 27 07:23:35 crc kubenswrapper[4706]: W1127 07:23:35.243087 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82d9ba46_8fc0_462c_bc77_056ef3252965.slice/crio-a971779c215b418f33527d7afe27d749aa1708e2f720768131b496d35a026ff3 WatchSource:0}: Error finding container a971779c215b418f33527d7afe27d749aa1708e2f720768131b496d35a026ff3: Status 404 returned error can't find the container with id a971779c215b418f33527d7afe27d749aa1708e2f720768131b496d35a026ff3 Nov 27 07:23:35 crc kubenswrapper[4706]: I1127 07:23:35.551018 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerID="fa2a6b2b83c7e6262e1e8885b82b53deb1a8f3687cc2e5205bd39b03c2dd89ce" exitCode=0 Nov 27 07:23:35 crc kubenswrapper[4706]: I1127 07:23:35.551066 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" event={"ID":"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee","Type":"ContainerDied","Data":"fa2a6b2b83c7e6262e1e8885b82b53deb1a8f3687cc2e5205bd39b03c2dd89ce"} Nov 27 07:23:35 crc kubenswrapper[4706]: I1127 07:23:35.555128 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" event={"ID":"82d9ba46-8fc0-462c-bc77-056ef3252965","Type":"ContainerStarted","Data":"a971779c215b418f33527d7afe27d749aa1708e2f720768131b496d35a026ff3"} Nov 27 07:23:36 crc kubenswrapper[4706]: I1127 07:23:36.244837 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:36 crc kubenswrapper[4706]: I1127 07:23:36.245167 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:36 crc kubenswrapper[4706]: I1127 07:23:36.299131 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:36 crc kubenswrapper[4706]: I1127 07:23:36.602428 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.261211 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-57d5z"] Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.264163 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.277562 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-57d5z"] Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.366889 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfmx9\" (UniqueName: \"kubernetes.io/projected/581cc62e-8174-4e0f-973f-96aad5109aa5-kube-api-access-hfmx9\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.367327 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-utilities\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.367402 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-catalog-content\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.468539 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfmx9\" (UniqueName: \"kubernetes.io/projected/581cc62e-8174-4e0f-973f-96aad5109aa5-kube-api-access-hfmx9\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.468626 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-utilities\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.468704 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-catalog-content\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.469310 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-catalog-content\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.469376 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-utilities\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.489451 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfmx9\" (UniqueName: \"kubernetes.io/projected/581cc62e-8174-4e0f-973f-96aad5109aa5-kube-api-access-hfmx9\") pod \"redhat-operators-57d5z\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.569975 4706 generic.go:334] "Generic (PLEG): container finished" podID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerID="f0883e863dee43c96eb401c4433a32dee80d1bbc6768e71ceff1a24ebcc6c5e6" exitCode=0 Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.570046 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" event={"ID":"82d9ba46-8fc0-462c-bc77-056ef3252965","Type":"ContainerDied","Data":"f0883e863dee43c96eb401c4433a32dee80d1bbc6768e71ceff1a24ebcc6c5e6"} Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.573931 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerID="98c4fb18458a1412f153c1aafd7ddb8c730c571c8130eeec548811699554f7b4" exitCode=0 Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.574049 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" event={"ID":"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee","Type":"ContainerDied","Data":"98c4fb18458a1412f153c1aafd7ddb8c730c571c8130eeec548811699554f7b4"} Nov 27 07:23:37 crc kubenswrapper[4706]: I1127 07:23:37.603660 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:38 crc kubenswrapper[4706]: I1127 07:23:38.037900 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-57d5z"] Nov 27 07:23:38 crc kubenswrapper[4706]: W1127 07:23:38.049478 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod581cc62e_8174_4e0f_973f_96aad5109aa5.slice/crio-0dc1d3718bd5e7f6118a3ef8c7f25bbba437b9dd37d0b56bf9a95c0e6ba71eed WatchSource:0}: Error finding container 0dc1d3718bd5e7f6118a3ef8c7f25bbba437b9dd37d0b56bf9a95c0e6ba71eed: Status 404 returned error can't find the container with id 0dc1d3718bd5e7f6118a3ef8c7f25bbba437b9dd37d0b56bf9a95c0e6ba71eed Nov 27 07:23:38 crc kubenswrapper[4706]: I1127 07:23:38.587601 4706 generic.go:334] "Generic (PLEG): container finished" podID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerID="886b0fc2c88748cfd1ccdec5fc7b36fbdababc8eab452e815f5485e83223ee94" exitCode=0 Nov 27 07:23:38 crc kubenswrapper[4706]: I1127 07:23:38.587756 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" event={"ID":"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee","Type":"ContainerDied","Data":"886b0fc2c88748cfd1ccdec5fc7b36fbdababc8eab452e815f5485e83223ee94"} Nov 27 07:23:38 crc kubenswrapper[4706]: I1127 07:23:38.594997 4706 generic.go:334] "Generic (PLEG): container finished" podID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerID="b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d" exitCode=0 Nov 27 07:23:38 crc kubenswrapper[4706]: I1127 07:23:38.595042 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57d5z" event={"ID":"581cc62e-8174-4e0f-973f-96aad5109aa5","Type":"ContainerDied","Data":"b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d"} Nov 27 07:23:38 crc kubenswrapper[4706]: I1127 07:23:38.595243 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57d5z" event={"ID":"581cc62e-8174-4e0f-973f-96aad5109aa5","Type":"ContainerStarted","Data":"0dc1d3718bd5e7f6118a3ef8c7f25bbba437b9dd37d0b56bf9a95c0e6ba71eed"} Nov 27 07:23:39 crc kubenswrapper[4706]: I1127 07:23:39.606421 4706 generic.go:334] "Generic (PLEG): container finished" podID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerID="40629e5c0d5be287f85f1b227900205c530324a791ae3e7abb26320003a8fc54" exitCode=0 Nov 27 07:23:39 crc kubenswrapper[4706]: I1127 07:23:39.606600 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" event={"ID":"82d9ba46-8fc0-462c-bc77-056ef3252965","Type":"ContainerDied","Data":"40629e5c0d5be287f85f1b227900205c530324a791ae3e7abb26320003a8fc54"} Nov 27 07:23:39 crc kubenswrapper[4706]: I1127 07:23:39.613215 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57d5z" event={"ID":"581cc62e-8174-4e0f-973f-96aad5109aa5","Type":"ContainerStarted","Data":"ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489"} Nov 27 07:23:39 crc kubenswrapper[4706]: I1127 07:23:39.933547 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.004002 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx2bq\" (UniqueName: \"kubernetes.io/projected/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-kube-api-access-tx2bq\") pod \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.004182 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-bundle\") pod \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.004251 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-util\") pod \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\" (UID: \"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee\") " Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.005187 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-bundle" (OuterVolumeSpecName: "bundle") pod "bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" (UID: "bf0b4f4f-059a-4c64-a75c-cffa8acd86ee"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.009903 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-kube-api-access-tx2bq" (OuterVolumeSpecName: "kube-api-access-tx2bq") pod "bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" (UID: "bf0b4f4f-059a-4c64-a75c-cffa8acd86ee"). InnerVolumeSpecName "kube-api-access-tx2bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.019814 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-util" (OuterVolumeSpecName: "util") pod "bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" (UID: "bf0b4f4f-059a-4c64-a75c-cffa8acd86ee"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.106305 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.106346 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.106358 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx2bq\" (UniqueName: \"kubernetes.io/projected/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee-kube-api-access-tx2bq\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.390128 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.390409 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.454398 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.621165 4706 generic.go:334] "Generic (PLEG): container finished" podID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerID="ac5b697565ab484fd78f13d3ba960e273a44b070383d109179a337f74091e088" exitCode=0 Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.621253 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" event={"ID":"82d9ba46-8fc0-462c-bc77-056ef3252965","Type":"ContainerDied","Data":"ac5b697565ab484fd78f13d3ba960e273a44b070383d109179a337f74091e088"} Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.623254 4706 generic.go:334] "Generic (PLEG): container finished" podID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerID="ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489" exitCode=0 Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.623339 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57d5z" event={"ID":"581cc62e-8174-4e0f-973f-96aad5109aa5","Type":"ContainerDied","Data":"ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489"} Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.626135 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.628651 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9" event={"ID":"bf0b4f4f-059a-4c64-a75c-cffa8acd86ee","Type":"ContainerDied","Data":"f33a904ac5f2541327d76425d5581734fb62c12a06cff11fda1baf6bfc6b3613"} Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.628706 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f33a904ac5f2541327d76425d5581734fb62c12a06cff11fda1baf6bfc6b3613" Nov 27 07:23:40 crc kubenswrapper[4706]: I1127 07:23:40.681455 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:41 crc kubenswrapper[4706]: I1127 07:23:41.633510 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57d5z" event={"ID":"581cc62e-8174-4e0f-973f-96aad5109aa5","Type":"ContainerStarted","Data":"17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162"} Nov 27 07:23:41 crc kubenswrapper[4706]: I1127 07:23:41.668454 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-57d5z" podStartSLOduration=2.198001729 podStartE2EDuration="4.668428678s" podCreationTimestamp="2025-11-27 07:23:37 +0000 UTC" firstStartedPulling="2025-11-27 07:23:38.596715575 +0000 UTC m=+902.486306385" lastFinishedPulling="2025-11-27 07:23:41.067142524 +0000 UTC m=+904.956733334" observedRunningTime="2025-11-27 07:23:41.662770139 +0000 UTC m=+905.552360959" watchObservedRunningTime="2025-11-27 07:23:41.668428678 +0000 UTC m=+905.558019518" Nov 27 07:23:41 crc kubenswrapper[4706]: I1127 07:23:41.933344 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.031796 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5gm4\" (UniqueName: \"kubernetes.io/projected/82d9ba46-8fc0-462c-bc77-056ef3252965-kube-api-access-w5gm4\") pod \"82d9ba46-8fc0-462c-bc77-056ef3252965\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.031904 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-bundle\") pod \"82d9ba46-8fc0-462c-bc77-056ef3252965\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.031947 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-util\") pod \"82d9ba46-8fc0-462c-bc77-056ef3252965\" (UID: \"82d9ba46-8fc0-462c-bc77-056ef3252965\") " Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.032704 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-bundle" (OuterVolumeSpecName: "bundle") pod "82d9ba46-8fc0-462c-bc77-056ef3252965" (UID: "82d9ba46-8fc0-462c-bc77-056ef3252965"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.042424 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82d9ba46-8fc0-462c-bc77-056ef3252965-kube-api-access-w5gm4" (OuterVolumeSpecName: "kube-api-access-w5gm4") pod "82d9ba46-8fc0-462c-bc77-056ef3252965" (UID: "82d9ba46-8fc0-462c-bc77-056ef3252965"). InnerVolumeSpecName "kube-api-access-w5gm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.045997 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-util" (OuterVolumeSpecName: "util") pod "82d9ba46-8fc0-462c-bc77-056ef3252965" (UID: "82d9ba46-8fc0-462c-bc77-056ef3252965"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.055125 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxkkm"] Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.055461 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cxkkm" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="registry-server" containerID="cri-o://525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422" gracePeriod=2 Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.133193 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.133245 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82d9ba46-8fc0-462c-bc77-056ef3252965-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.133259 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5gm4\" (UniqueName: \"kubernetes.io/projected/82d9ba46-8fc0-462c-bc77-056ef3252965-kube-api-access-w5gm4\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.434877 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.537206 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmhq8\" (UniqueName: \"kubernetes.io/projected/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-kube-api-access-tmhq8\") pod \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.537381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-catalog-content\") pod \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.537443 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-utilities\") pod \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\" (UID: \"33052f7e-58dc-472e-8ffd-2e6fd9c16c27\") " Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.538062 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-utilities" (OuterVolumeSpecName: "utilities") pod "33052f7e-58dc-472e-8ffd-2e6fd9c16c27" (UID: "33052f7e-58dc-472e-8ffd-2e6fd9c16c27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.538305 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.545936 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-kube-api-access-tmhq8" (OuterVolumeSpecName: "kube-api-access-tmhq8") pod "33052f7e-58dc-472e-8ffd-2e6fd9c16c27" (UID: "33052f7e-58dc-472e-8ffd-2e6fd9c16c27"). InnerVolumeSpecName "kube-api-access-tmhq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.554279 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33052f7e-58dc-472e-8ffd-2e6fd9c16c27" (UID: "33052f7e-58dc-472e-8ffd-2e6fd9c16c27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.639296 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmhq8\" (UniqueName: \"kubernetes.io/projected/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-kube-api-access-tmhq8\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.639326 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33052f7e-58dc-472e-8ffd-2e6fd9c16c27-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.642993 4706 generic.go:334] "Generic (PLEG): container finished" podID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerID="525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422" exitCode=0 Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.643053 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxkkm" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.643083 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxkkm" event={"ID":"33052f7e-58dc-472e-8ffd-2e6fd9c16c27","Type":"ContainerDied","Data":"525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422"} Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.643134 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxkkm" event={"ID":"33052f7e-58dc-472e-8ffd-2e6fd9c16c27","Type":"ContainerDied","Data":"be8d58a05bad5371a9fe265d903195f7d59d9cfe0f130de298f1d2f317e5e031"} Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.655173 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.655166 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6" event={"ID":"82d9ba46-8fc0-462c-bc77-056ef3252965","Type":"ContainerDied","Data":"a971779c215b418f33527d7afe27d749aa1708e2f720768131b496d35a026ff3"} Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.655382 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a971779c215b418f33527d7afe27d749aa1708e2f720768131b496d35a026ff3" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.643181 4706 scope.go:117] "RemoveContainer" containerID="525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.689968 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxkkm"] Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.697596 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxkkm"] Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.701938 4706 scope.go:117] "RemoveContainer" containerID="8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.719575 4706 scope.go:117] "RemoveContainer" containerID="d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.744467 4706 scope.go:117] "RemoveContainer" containerID="525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422" Nov 27 07:23:42 crc kubenswrapper[4706]: E1127 07:23:42.745643 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422\": container with ID starting with 525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422 not found: ID does not exist" containerID="525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.745695 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422"} err="failed to get container status \"525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422\": rpc error: code = NotFound desc = could not find container \"525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422\": container with ID starting with 525e659d4e4652e5c6e50d0e73a1f852296efd45cbdda4ca9518a25871c44422 not found: ID does not exist" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.745730 4706 scope.go:117] "RemoveContainer" containerID="8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6" Nov 27 07:23:42 crc kubenswrapper[4706]: E1127 07:23:42.746995 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6\": container with ID starting with 8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6 not found: ID does not exist" containerID="8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.747021 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6"} err="failed to get container status \"8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6\": rpc error: code = NotFound desc = could not find container \"8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6\": container with ID starting with 8b85b876e67501eaa9cce806a128c45e7c533e3eb7e9badc5c85c5396169dbb6 not found: ID does not exist" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.747034 4706 scope.go:117] "RemoveContainer" containerID="d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9" Nov 27 07:23:42 crc kubenswrapper[4706]: E1127 07:23:42.750407 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9\": container with ID starting with d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9 not found: ID does not exist" containerID="d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.750445 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9"} err="failed to get container status \"d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9\": rpc error: code = NotFound desc = could not find container \"d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9\": container with ID starting with d3ff035060c4e12ed136efa9088f724cfa9c5d53096be1df1f627cca22d77cf9 not found: ID does not exist" Nov 27 07:23:42 crc kubenswrapper[4706]: I1127 07:23:42.806902 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" path="/var/lib/kubelet/pods/33052f7e-58dc-472e-8ffd-2e6fd9c16c27/volumes" Nov 27 07:23:45 crc kubenswrapper[4706]: I1127 07:23:45.177847 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:23:45 crc kubenswrapper[4706]: I1127 07:23:45.178175 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:23:47 crc kubenswrapper[4706]: I1127 07:23:47.603940 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:47 crc kubenswrapper[4706]: I1127 07:23:47.604836 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:47 crc kubenswrapper[4706]: I1127 07:23:47.668124 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:47 crc kubenswrapper[4706]: I1127 07:23:47.741746 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.257575 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pblwf"] Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.257846 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pblwf" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="registry-server" containerID="cri-o://93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978" gracePeriod=2 Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.650298 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.709360 4706 generic.go:334] "Generic (PLEG): container finished" podID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerID="93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978" exitCode=0 Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.709415 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pblwf" event={"ID":"49808ffe-83fb-4aa7-8eae-8debb8354286","Type":"ContainerDied","Data":"93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978"} Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.709443 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pblwf" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.709473 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pblwf" event={"ID":"49808ffe-83fb-4aa7-8eae-8debb8354286","Type":"ContainerDied","Data":"da73a168cb03772a3c5cd9d60e161d6aaedcc4eb2b3795736ec3e04872ff389e"} Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.709489 4706 scope.go:117] "RemoveContainer" containerID="93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.723591 4706 scope.go:117] "RemoveContainer" containerID="9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.729159 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-catalog-content\") pod \"49808ffe-83fb-4aa7-8eae-8debb8354286\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.729286 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnn4k\" (UniqueName: \"kubernetes.io/projected/49808ffe-83fb-4aa7-8eae-8debb8354286-kube-api-access-lnn4k\") pod \"49808ffe-83fb-4aa7-8eae-8debb8354286\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.729415 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-utilities\") pod \"49808ffe-83fb-4aa7-8eae-8debb8354286\" (UID: \"49808ffe-83fb-4aa7-8eae-8debb8354286\") " Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.730396 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-utilities" (OuterVolumeSpecName: "utilities") pod "49808ffe-83fb-4aa7-8eae-8debb8354286" (UID: "49808ffe-83fb-4aa7-8eae-8debb8354286"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.736335 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49808ffe-83fb-4aa7-8eae-8debb8354286-kube-api-access-lnn4k" (OuterVolumeSpecName: "kube-api-access-lnn4k") pod "49808ffe-83fb-4aa7-8eae-8debb8354286" (UID: "49808ffe-83fb-4aa7-8eae-8debb8354286"). InnerVolumeSpecName "kube-api-access-lnn4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.741575 4706 scope.go:117] "RemoveContainer" containerID="17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.784032 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49808ffe-83fb-4aa7-8eae-8debb8354286" (UID: "49808ffe-83fb-4aa7-8eae-8debb8354286"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.785329 4706 scope.go:117] "RemoveContainer" containerID="93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978" Nov 27 07:23:48 crc kubenswrapper[4706]: E1127 07:23:48.785666 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978\": container with ID starting with 93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978 not found: ID does not exist" containerID="93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.785710 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978"} err="failed to get container status \"93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978\": rpc error: code = NotFound desc = could not find container \"93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978\": container with ID starting with 93e0ed25c8f461b7f6c42105d370a8526f8526f0886b10798d26b9a3d48c4978 not found: ID does not exist" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.785738 4706 scope.go:117] "RemoveContainer" containerID="9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc" Nov 27 07:23:48 crc kubenswrapper[4706]: E1127 07:23:48.785976 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc\": container with ID starting with 9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc not found: ID does not exist" containerID="9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.786009 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc"} err="failed to get container status \"9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc\": rpc error: code = NotFound desc = could not find container \"9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc\": container with ID starting with 9714e1b9b03ed2ad418b1f793c94cc8ec978aefbeab19e76cbfc46596212ccfc not found: ID does not exist" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.786031 4706 scope.go:117] "RemoveContainer" containerID="17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d" Nov 27 07:23:48 crc kubenswrapper[4706]: E1127 07:23:48.786306 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d\": container with ID starting with 17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d not found: ID does not exist" containerID="17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.786331 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d"} err="failed to get container status \"17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d\": rpc error: code = NotFound desc = could not find container \"17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d\": container with ID starting with 17b9ca223e54104f032c328ee7c52762f8b1acaed6d4b84fd6b5b2279f5d445d not found: ID does not exist" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.830555 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.830585 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49808ffe-83fb-4aa7-8eae-8debb8354286-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:48 crc kubenswrapper[4706]: I1127 07:23:48.830596 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnn4k\" (UniqueName: \"kubernetes.io/projected/49808ffe-83fb-4aa7-8eae-8debb8354286-kube-api-access-lnn4k\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:49 crc kubenswrapper[4706]: I1127 07:23:49.027909 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pblwf"] Nov 27 07:23:49 crc kubenswrapper[4706]: I1127 07:23:49.034177 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pblwf"] Nov 27 07:23:50 crc kubenswrapper[4706]: I1127 07:23:50.784518 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" path="/var/lib/kubelet/pods/49808ffe-83fb-4aa7-8eae-8debb8354286/volumes" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.446958 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q"] Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447565 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="registry-server" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447578 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="registry-server" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447597 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerName="util" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447603 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerName="util" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447611 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerName="extract" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447617 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerName="extract" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447626 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerName="pull" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447632 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerName="pull" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447640 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="extract-utilities" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447647 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="extract-utilities" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447661 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="registry-server" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447666 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="registry-server" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447676 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerName="extract" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447681 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerName="extract" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447690 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerName="pull" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447696 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerName="pull" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447705 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerName="util" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447710 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerName="util" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447720 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="extract-content" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447725 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="extract-content" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447732 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="extract-content" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447739 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="extract-content" Nov 27 07:23:52 crc kubenswrapper[4706]: E1127 07:23:52.447746 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="extract-utilities" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447754 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="extract-utilities" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447887 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" containerName="extract" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447898 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="49808ffe-83fb-4aa7-8eae-8debb8354286" containerName="registry-server" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447905 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="33052f7e-58dc-472e-8ffd-2e6fd9c16c27" containerName="registry-server" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.447914 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d9ba46-8fc0-462c-bc77-056ef3252965" containerName="extract" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.448353 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.450190 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.453214 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-bctd8" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.477813 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q"] Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.582103 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-apiservice-cert\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.582151 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpq6d\" (UniqueName: \"kubernetes.io/projected/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-kube-api-access-dpq6d\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.582196 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-webhook-cert\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.683970 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-apiservice-cert\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.684025 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpq6d\" (UniqueName: \"kubernetes.io/projected/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-kube-api-access-dpq6d\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.684089 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-webhook-cert\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.696244 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-webhook-cert\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.696671 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-apiservice-cert\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.705470 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpq6d\" (UniqueName: \"kubernetes.io/projected/0961e9eb-dadf-4e6a-bbbb-34cd06d1b683-kube-api-access-dpq6d\") pod \"horizon-operator-controller-manager-54fdcb5fdb-dn97q\" (UID: \"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683\") " pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:52 crc kubenswrapper[4706]: I1127 07:23:52.765742 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:53 crc kubenswrapper[4706]: I1127 07:23:53.038655 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q"] Nov 27 07:23:53 crc kubenswrapper[4706]: I1127 07:23:53.739697 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" event={"ID":"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683","Type":"ContainerStarted","Data":"d9b7632265b8f456cd43e488a2f2b2335f984d79a8f4d81f0e1d3d75655e4d15"} Nov 27 07:23:55 crc kubenswrapper[4706]: I1127 07:23:55.754449 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" event={"ID":"0961e9eb-dadf-4e6a-bbbb-34cd06d1b683","Type":"ContainerStarted","Data":"4ab5b777460552bb426b412302eb8be92c3d4d07766d4cf37f398656ed1fc7d0"} Nov 27 07:23:55 crc kubenswrapper[4706]: I1127 07:23:55.755087 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:23:55 crc kubenswrapper[4706]: I1127 07:23:55.774425 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" podStartSLOduration=1.803632055 podStartE2EDuration="3.774401496s" podCreationTimestamp="2025-11-27 07:23:52 +0000 UTC" firstStartedPulling="2025-11-27 07:23:53.046106594 +0000 UTC m=+916.935697404" lastFinishedPulling="2025-11-27 07:23:55.016876035 +0000 UTC m=+918.906466845" observedRunningTime="2025-11-27 07:23:55.768674435 +0000 UTC m=+919.658265255" watchObservedRunningTime="2025-11-27 07:23:55.774401496 +0000 UTC m=+919.663992306" Nov 27 07:23:56 crc kubenswrapper[4706]: I1127 07:23:56.861105 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-57d5z"] Nov 27 07:23:56 crc kubenswrapper[4706]: I1127 07:23:56.861334 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-57d5z" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="registry-server" containerID="cri-o://17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162" gracePeriod=2 Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.222366 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.365366 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-utilities\") pod \"581cc62e-8174-4e0f-973f-96aad5109aa5\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.365430 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-catalog-content\") pod \"581cc62e-8174-4e0f-973f-96aad5109aa5\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.365458 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfmx9\" (UniqueName: \"kubernetes.io/projected/581cc62e-8174-4e0f-973f-96aad5109aa5-kube-api-access-hfmx9\") pod \"581cc62e-8174-4e0f-973f-96aad5109aa5\" (UID: \"581cc62e-8174-4e0f-973f-96aad5109aa5\") " Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.366253 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-utilities" (OuterVolumeSpecName: "utilities") pod "581cc62e-8174-4e0f-973f-96aad5109aa5" (UID: "581cc62e-8174-4e0f-973f-96aad5109aa5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.375808 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/581cc62e-8174-4e0f-973f-96aad5109aa5-kube-api-access-hfmx9" (OuterVolumeSpecName: "kube-api-access-hfmx9") pod "581cc62e-8174-4e0f-973f-96aad5109aa5" (UID: "581cc62e-8174-4e0f-973f-96aad5109aa5"). InnerVolumeSpecName "kube-api-access-hfmx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.452494 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "581cc62e-8174-4e0f-973f-96aad5109aa5" (UID: "581cc62e-8174-4e0f-973f-96aad5109aa5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.467121 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.467160 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfmx9\" (UniqueName: \"kubernetes.io/projected/581cc62e-8174-4e0f-973f-96aad5109aa5-kube-api-access-hfmx9\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.467175 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581cc62e-8174-4e0f-973f-96aad5109aa5-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.768408 4706 generic.go:334] "Generic (PLEG): container finished" podID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerID="17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162" exitCode=0 Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.768459 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57d5z" event={"ID":"581cc62e-8174-4e0f-973f-96aad5109aa5","Type":"ContainerDied","Data":"17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162"} Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.768492 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57d5z" event={"ID":"581cc62e-8174-4e0f-973f-96aad5109aa5","Type":"ContainerDied","Data":"0dc1d3718bd5e7f6118a3ef8c7f25bbba437b9dd37d0b56bf9a95c0e6ba71eed"} Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.768513 4706 scope.go:117] "RemoveContainer" containerID="17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.768513 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57d5z" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.817416 4706 scope.go:117] "RemoveContainer" containerID="ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.840547 4706 scope.go:117] "RemoveContainer" containerID="b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.850579 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-57d5z"] Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.858041 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-57d5z"] Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.865291 4706 scope.go:117] "RemoveContainer" containerID="17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162" Nov 27 07:23:57 crc kubenswrapper[4706]: E1127 07:23:57.865684 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162\": container with ID starting with 17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162 not found: ID does not exist" containerID="17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.865710 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162"} err="failed to get container status \"17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162\": rpc error: code = NotFound desc = could not find container \"17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162\": container with ID starting with 17d052fbf7694b6df88aabb3fa434945252ac06db7988ed149a69c8e63eef162 not found: ID does not exist" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.865731 4706 scope.go:117] "RemoveContainer" containerID="ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489" Nov 27 07:23:57 crc kubenswrapper[4706]: E1127 07:23:57.865952 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489\": container with ID starting with ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489 not found: ID does not exist" containerID="ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.865969 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489"} err="failed to get container status \"ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489\": rpc error: code = NotFound desc = could not find container \"ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489\": container with ID starting with ed7128e1eef5c8175af8ea111f72ddb2fc24c86d45f3b1ab4ad095fd82d93489 not found: ID does not exist" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.865981 4706 scope.go:117] "RemoveContainer" containerID="b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d" Nov 27 07:23:57 crc kubenswrapper[4706]: E1127 07:23:57.866213 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d\": container with ID starting with b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d not found: ID does not exist" containerID="b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d" Nov 27 07:23:57 crc kubenswrapper[4706]: I1127 07:23:57.866240 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d"} err="failed to get container status \"b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d\": rpc error: code = NotFound desc = could not find container \"b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d\": container with ID starting with b5a01a1af9897e02e0b0dca3a0fd782cc1e6dfeaae1b64e93562374467f9315d not found: ID does not exist" Nov 27 07:23:58 crc kubenswrapper[4706]: I1127 07:23:58.783431 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" path="/var/lib/kubelet/pods/581cc62e-8174-4e0f-973f-96aad5109aa5/volumes" Nov 27 07:23:59 crc kubenswrapper[4706]: I1127 07:23:59.418267 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.277884 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm"] Nov 27 07:24:00 crc kubenswrapper[4706]: E1127 07:24:00.278406 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="registry-server" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.278508 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="registry-server" Nov 27 07:24:00 crc kubenswrapper[4706]: E1127 07:24:00.278586 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="extract-content" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.278658 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="extract-content" Nov 27 07:24:00 crc kubenswrapper[4706]: E1127 07:24:00.278725 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="extract-utilities" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.278780 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="extract-utilities" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.278966 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="581cc62e-8174-4e0f-973f-96aad5109aa5" containerName="registry-server" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.279471 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.282460 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-7czbv" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.282829 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.297835 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm"] Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.421300 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-webhook-cert\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.421364 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck4wh\" (UniqueName: \"kubernetes.io/projected/4e65f1dd-27d0-4445-b716-7e656f7c85c6-kube-api-access-ck4wh\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.421494 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-apiservice-cert\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.522250 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-webhook-cert\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.522334 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck4wh\" (UniqueName: \"kubernetes.io/projected/4e65f1dd-27d0-4445-b716-7e656f7c85c6-kube-api-access-ck4wh\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.522389 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-apiservice-cert\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.530422 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-apiservice-cert\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.532090 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-webhook-cert\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.541252 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck4wh\" (UniqueName: \"kubernetes.io/projected/4e65f1dd-27d0-4445-b716-7e656f7c85c6-kube-api-access-ck4wh\") pod \"swift-operator-controller-manager-7c4fb59c48-m6nmm\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:00 crc kubenswrapper[4706]: I1127 07:24:00.645672 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:01 crc kubenswrapper[4706]: I1127 07:24:01.055042 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm"] Nov 27 07:24:01 crc kubenswrapper[4706]: I1127 07:24:01.070596 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 07:24:01 crc kubenswrapper[4706]: I1127 07:24:01.812579 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" event={"ID":"4e65f1dd-27d0-4445-b716-7e656f7c85c6","Type":"ContainerStarted","Data":"d54530a6e59ec41d30f64ff40e75fc5520b755e0635f6a5aee1de2a3c1fe6f1b"} Nov 27 07:24:02 crc kubenswrapper[4706]: I1127 07:24:02.771308 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54fdcb5fdb-dn97q" Nov 27 07:24:03 crc kubenswrapper[4706]: I1127 07:24:03.828379 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" event={"ID":"4e65f1dd-27d0-4445-b716-7e656f7c85c6","Type":"ContainerStarted","Data":"76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57"} Nov 27 07:24:03 crc kubenswrapper[4706]: I1127 07:24:03.828731 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:03 crc kubenswrapper[4706]: I1127 07:24:03.857091 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" podStartSLOduration=1.942811813 podStartE2EDuration="3.857066912s" podCreationTimestamp="2025-11-27 07:24:00 +0000 UTC" firstStartedPulling="2025-11-27 07:24:01.070282229 +0000 UTC m=+924.959873039" lastFinishedPulling="2025-11-27 07:24:02.984537338 +0000 UTC m=+926.874128138" observedRunningTime="2025-11-27 07:24:03.850791924 +0000 UTC m=+927.740382744" watchObservedRunningTime="2025-11-27 07:24:03.857066912 +0000 UTC m=+927.746657722" Nov 27 07:24:10 crc kubenswrapper[4706]: I1127 07:24:10.649160 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.178154 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.178617 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.178670 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.179509 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db4ade098e79285766f3d1f1fb101ceeccbe33f1280ddb92a2252c3b8d5a2939"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.179563 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://db4ade098e79285766f3d1f1fb101ceeccbe33f1280ddb92a2252c3b8d5a2939" gracePeriod=600 Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.908680 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="db4ade098e79285766f3d1f1fb101ceeccbe33f1280ddb92a2252c3b8d5a2939" exitCode=0 Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.908761 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"db4ade098e79285766f3d1f1fb101ceeccbe33f1280ddb92a2252c3b8d5a2939"} Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.909071 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"d7703c2342c25e6f9c346c6d641fa2f157694340cd2c97e78820343e2f8c3744"} Nov 27 07:24:15 crc kubenswrapper[4706]: I1127 07:24:15.909116 4706 scope.go:117] "RemoveContainer" containerID="4cee0d2f37ffae96c2927fba888af9c607b55ddd51e1dd0044435dc5bccfed27" Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.062860 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-54852"] Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.064228 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-54852" Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.066653 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-6shtc" Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.072977 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-54852"] Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.160294 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2g6l\" (UniqueName: \"kubernetes.io/projected/55821f55-ec95-4d85-bc6d-c7aeef199f3c-kube-api-access-n2g6l\") pod \"glance-operator-index-54852\" (UID: \"55821f55-ec95-4d85-bc6d-c7aeef199f3c\") " pod="openstack-operators/glance-operator-index-54852" Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.261269 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2g6l\" (UniqueName: \"kubernetes.io/projected/55821f55-ec95-4d85-bc6d-c7aeef199f3c-kube-api-access-n2g6l\") pod \"glance-operator-index-54852\" (UID: \"55821f55-ec95-4d85-bc6d-c7aeef199f3c\") " pod="openstack-operators/glance-operator-index-54852" Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.280272 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2g6l\" (UniqueName: \"kubernetes.io/projected/55821f55-ec95-4d85-bc6d-c7aeef199f3c-kube-api-access-n2g6l\") pod \"glance-operator-index-54852\" (UID: \"55821f55-ec95-4d85-bc6d-c7aeef199f3c\") " pod="openstack-operators/glance-operator-index-54852" Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.379278 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-54852" Nov 27 07:24:16 crc kubenswrapper[4706]: I1127 07:24:16.977188 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-54852"] Nov 27 07:24:16 crc kubenswrapper[4706]: W1127 07:24:16.990283 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55821f55_ec95_4d85_bc6d_c7aeef199f3c.slice/crio-7cbeba730b5346ec21ea3bf94136dae208129f9d7c0c8e80dcd79cdab9d0d4c1 WatchSource:0}: Error finding container 7cbeba730b5346ec21ea3bf94136dae208129f9d7c0c8e80dcd79cdab9d0d4c1: Status 404 returned error can't find the container with id 7cbeba730b5346ec21ea3bf94136dae208129f9d7c0c8e80dcd79cdab9d0d4c1 Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.547610 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.552577 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.557432 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.557447 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.557579 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.558859 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-rgn8t" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.571957 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.678878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.679165 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.679231 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-cache\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.679270 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cszr\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-kube-api-access-5cszr\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.679391 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-lock\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.780584 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.780625 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-cache\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.780645 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cszr\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-kube-api-access-5cszr\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.780670 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-lock\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.780710 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: E1127 07:24:17.780844 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:17 crc kubenswrapper[4706]: E1127 07:24:17.780857 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 27 07:24:17 crc kubenswrapper[4706]: E1127 07:24:17.780904 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift podName:85c33b82-9ee0-4e79-82ee-f9de8a9bfefd nodeName:}" failed. No retries permitted until 2025-11-27 07:24:18.280886558 +0000 UTC m=+942.170477368 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift") pod "swift-storage-0" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd") : configmap "swift-ring-files" not found Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.780963 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.781039 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-cache\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.781146 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-lock\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.799831 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cszr\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-kube-api-access-5cszr\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.805958 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:17 crc kubenswrapper[4706]: I1127 07:24:17.926720 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-54852" event={"ID":"55821f55-ec95-4d85-bc6d-c7aeef199f3c","Type":"ContainerStarted","Data":"7cbeba730b5346ec21ea3bf94136dae208129f9d7c0c8e80dcd79cdab9d0d4c1"} Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.029123 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-2zfbf"] Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.030099 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.031738 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.032539 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.032822 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.050225 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-2zfbf"] Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.085487 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-scripts\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.085534 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-dispersionconf\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.085626 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-swiftconf\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.085649 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5dgv\" (UniqueName: \"kubernetes.io/projected/525e7753-926d-429b-bbe0-f45e95cf80c3-kube-api-access-k5dgv\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.085678 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-ring-data-devices\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.085715 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/525e7753-926d-429b-bbe0-f45e95cf80c3-etc-swift\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.187705 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-swiftconf\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.187756 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5dgv\" (UniqueName: \"kubernetes.io/projected/525e7753-926d-429b-bbe0-f45e95cf80c3-kube-api-access-k5dgv\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.187779 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-ring-data-devices\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.187804 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/525e7753-926d-429b-bbe0-f45e95cf80c3-etc-swift\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.187851 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-scripts\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.187870 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-dispersionconf\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.188849 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-ring-data-devices\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.188871 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/525e7753-926d-429b-bbe0-f45e95cf80c3-etc-swift\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.189947 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-scripts\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.196915 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-swiftconf\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.206053 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-dispersionconf\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.224469 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5dgv\" (UniqueName: \"kubernetes.io/projected/525e7753-926d-429b-bbe0-f45e95cf80c3-kube-api-access-k5dgv\") pod \"swift-ring-rebalance-2zfbf\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.289151 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:18 crc kubenswrapper[4706]: E1127 07:24:18.289356 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:18 crc kubenswrapper[4706]: E1127 07:24:18.289371 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 27 07:24:18 crc kubenswrapper[4706]: E1127 07:24:18.289414 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift podName:85c33b82-9ee0-4e79-82ee-f9de8a9bfefd nodeName:}" failed. No retries permitted until 2025-11-27 07:24:19.289399867 +0000 UTC m=+943.178990677 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift") pod "swift-storage-0" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd") : configmap "swift-ring-files" not found Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.344508 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.346298 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs"] Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.347346 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.353139 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs"] Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.491131 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-log-httpd\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.491181 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-run-httpd\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.491199 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.491226 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trfrj\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-kube-api-access-trfrj\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.491372 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-config-data\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.592841 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-config-data\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.592934 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-log-httpd\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.592961 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-run-httpd\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.592980 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.593008 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trfrj\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-kube-api-access-trfrj\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.593728 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-log-httpd\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.593793 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-run-httpd\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: E1127 07:24:18.593860 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:18 crc kubenswrapper[4706]: E1127 07:24:18.594083 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs: configmap "swift-ring-files" not found Nov 27 07:24:18 crc kubenswrapper[4706]: E1127 07:24:18.594126 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift podName:64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76 nodeName:}" failed. No retries permitted until 2025-11-27 07:24:19.094111616 +0000 UTC m=+942.983702426 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift") pod "swift-proxy-6bd58cfcf7-jd2hs" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76") : configmap "swift-ring-files" not found Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.599609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-config-data\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.610652 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trfrj\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-kube-api-access-trfrj\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.934027 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-54852" event={"ID":"55821f55-ec95-4d85-bc6d-c7aeef199f3c","Type":"ContainerStarted","Data":"b1e37dc1bac20e79e12376c3887c3fd4864cf2f01a71906618d02240b63cc3ce"} Nov 27 07:24:18 crc kubenswrapper[4706]: I1127 07:24:18.957186 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-54852" podStartSLOduration=1.291013827 podStartE2EDuration="2.957168881s" podCreationTimestamp="2025-11-27 07:24:16 +0000 UTC" firstStartedPulling="2025-11-27 07:24:16.992582006 +0000 UTC m=+940.882172816" lastFinishedPulling="2025-11-27 07:24:18.65873706 +0000 UTC m=+942.548327870" observedRunningTime="2025-11-27 07:24:18.953019567 +0000 UTC m=+942.842610377" watchObservedRunningTime="2025-11-27 07:24:18.957168881 +0000 UTC m=+942.846759691" Nov 27 07:24:19 crc kubenswrapper[4706]: I1127 07:24:19.024262 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-2zfbf"] Nov 27 07:24:19 crc kubenswrapper[4706]: I1127 07:24:19.119036 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:19 crc kubenswrapper[4706]: E1127 07:24:19.119297 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:19 crc kubenswrapper[4706]: E1127 07:24:19.119426 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs: configmap "swift-ring-files" not found Nov 27 07:24:19 crc kubenswrapper[4706]: E1127 07:24:19.119514 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift podName:64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76 nodeName:}" failed. No retries permitted until 2025-11-27 07:24:20.119490769 +0000 UTC m=+944.009081589 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift") pod "swift-proxy-6bd58cfcf7-jd2hs" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76") : configmap "swift-ring-files" not found Nov 27 07:24:19 crc kubenswrapper[4706]: I1127 07:24:19.322958 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:19 crc kubenswrapper[4706]: E1127 07:24:19.323205 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:19 crc kubenswrapper[4706]: E1127 07:24:19.323274 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 27 07:24:19 crc kubenswrapper[4706]: E1127 07:24:19.323347 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift podName:85c33b82-9ee0-4e79-82ee-f9de8a9bfefd nodeName:}" failed. No retries permitted until 2025-11-27 07:24:21.323322629 +0000 UTC m=+945.212913449 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift") pod "swift-storage-0" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd") : configmap "swift-ring-files" not found Nov 27 07:24:19 crc kubenswrapper[4706]: I1127 07:24:19.943713 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" event={"ID":"525e7753-926d-429b-bbe0-f45e95cf80c3","Type":"ContainerStarted","Data":"4a7c3b61ca649e490effdcfa590ba387faff7124c8b79bd0e2fed354229d0092"} Nov 27 07:24:20 crc kubenswrapper[4706]: I1127 07:24:20.135117 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:20 crc kubenswrapper[4706]: E1127 07:24:20.135301 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:20 crc kubenswrapper[4706]: E1127 07:24:20.135329 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs: configmap "swift-ring-files" not found Nov 27 07:24:20 crc kubenswrapper[4706]: E1127 07:24:20.135410 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift podName:64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76 nodeName:}" failed. No retries permitted until 2025-11-27 07:24:22.135362892 +0000 UTC m=+946.024953702 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift") pod "swift-proxy-6bd58cfcf7-jd2hs" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76") : configmap "swift-ring-files" not found Nov 27 07:24:20 crc kubenswrapper[4706]: I1127 07:24:20.258787 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-54852"] Nov 27 07:24:20 crc kubenswrapper[4706]: I1127 07:24:20.860031 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-q6g28"] Nov 27 07:24:20 crc kubenswrapper[4706]: I1127 07:24:20.861163 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:20 crc kubenswrapper[4706]: I1127 07:24:20.869478 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-q6g28"] Nov 27 07:24:20 crc kubenswrapper[4706]: I1127 07:24:20.947915 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rfwv\" (UniqueName: \"kubernetes.io/projected/b1511901-0bd6-44e5-a199-c99065ee8138-kube-api-access-8rfwv\") pod \"glance-operator-index-q6g28\" (UID: \"b1511901-0bd6-44e5-a199-c99065ee8138\") " pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:20 crc kubenswrapper[4706]: I1127 07:24:20.953532 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-54852" podUID="55821f55-ec95-4d85-bc6d-c7aeef199f3c" containerName="registry-server" containerID="cri-o://b1e37dc1bac20e79e12376c3887c3fd4864cf2f01a71906618d02240b63cc3ce" gracePeriod=2 Nov 27 07:24:21 crc kubenswrapper[4706]: I1127 07:24:21.049871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rfwv\" (UniqueName: \"kubernetes.io/projected/b1511901-0bd6-44e5-a199-c99065ee8138-kube-api-access-8rfwv\") pod \"glance-operator-index-q6g28\" (UID: \"b1511901-0bd6-44e5-a199-c99065ee8138\") " pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:21 crc kubenswrapper[4706]: I1127 07:24:21.095035 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rfwv\" (UniqueName: \"kubernetes.io/projected/b1511901-0bd6-44e5-a199-c99065ee8138-kube-api-access-8rfwv\") pod \"glance-operator-index-q6g28\" (UID: \"b1511901-0bd6-44e5-a199-c99065ee8138\") " pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:21 crc kubenswrapper[4706]: I1127 07:24:21.190038 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:21 crc kubenswrapper[4706]: I1127 07:24:21.355756 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:21 crc kubenswrapper[4706]: E1127 07:24:21.355928 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:21 crc kubenswrapper[4706]: E1127 07:24:21.355944 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 27 07:24:21 crc kubenswrapper[4706]: E1127 07:24:21.355997 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift podName:85c33b82-9ee0-4e79-82ee-f9de8a9bfefd nodeName:}" failed. No retries permitted until 2025-11-27 07:24:25.355977462 +0000 UTC m=+949.245568282 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift") pod "swift-storage-0" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd") : configmap "swift-ring-files" not found Nov 27 07:24:21 crc kubenswrapper[4706]: I1127 07:24:21.962152 4706 generic.go:334] "Generic (PLEG): container finished" podID="55821f55-ec95-4d85-bc6d-c7aeef199f3c" containerID="b1e37dc1bac20e79e12376c3887c3fd4864cf2f01a71906618d02240b63cc3ce" exitCode=0 Nov 27 07:24:21 crc kubenswrapper[4706]: I1127 07:24:21.962195 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-54852" event={"ID":"55821f55-ec95-4d85-bc6d-c7aeef199f3c","Type":"ContainerDied","Data":"b1e37dc1bac20e79e12376c3887c3fd4864cf2f01a71906618d02240b63cc3ce"} Nov 27 07:24:22 crc kubenswrapper[4706]: I1127 07:24:22.173682 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:22 crc kubenswrapper[4706]: E1127 07:24:22.173850 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:22 crc kubenswrapper[4706]: E1127 07:24:22.173882 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs: configmap "swift-ring-files" not found Nov 27 07:24:22 crc kubenswrapper[4706]: E1127 07:24:22.173946 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift podName:64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76 nodeName:}" failed. No retries permitted until 2025-11-27 07:24:26.173926071 +0000 UTC m=+950.063516881 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift") pod "swift-proxy-6bd58cfcf7-jd2hs" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76") : configmap "swift-ring-files" not found Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.165586 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-54852" Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.219510 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2g6l\" (UniqueName: \"kubernetes.io/projected/55821f55-ec95-4d85-bc6d-c7aeef199f3c-kube-api-access-n2g6l\") pod \"55821f55-ec95-4d85-bc6d-c7aeef199f3c\" (UID: \"55821f55-ec95-4d85-bc6d-c7aeef199f3c\") " Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.224429 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55821f55-ec95-4d85-bc6d-c7aeef199f3c-kube-api-access-n2g6l" (OuterVolumeSpecName: "kube-api-access-n2g6l") pod "55821f55-ec95-4d85-bc6d-c7aeef199f3c" (UID: "55821f55-ec95-4d85-bc6d-c7aeef199f3c"). InnerVolumeSpecName "kube-api-access-n2g6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.321665 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2g6l\" (UniqueName: \"kubernetes.io/projected/55821f55-ec95-4d85-bc6d-c7aeef199f3c-kube-api-access-n2g6l\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.422821 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:25 crc kubenswrapper[4706]: E1127 07:24:25.423006 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:25 crc kubenswrapper[4706]: E1127 07:24:25.423039 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 27 07:24:25 crc kubenswrapper[4706]: E1127 07:24:25.423090 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift podName:85c33b82-9ee0-4e79-82ee-f9de8a9bfefd nodeName:}" failed. No retries permitted until 2025-11-27 07:24:33.42307408 +0000 UTC m=+957.312664890 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift") pod "swift-storage-0" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd") : configmap "swift-ring-files" not found Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.424206 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-q6g28"] Nov 27 07:24:25 crc kubenswrapper[4706]: W1127 07:24:25.431969 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1511901_0bd6_44e5_a199_c99065ee8138.slice/crio-cb8d0d0c67f1d975d677ac13c2788ffe661f3101019ee63df86c0dbe1bfce76a WatchSource:0}: Error finding container cb8d0d0c67f1d975d677ac13c2788ffe661f3101019ee63df86c0dbe1bfce76a: Status 404 returned error can't find the container with id cb8d0d0c67f1d975d677ac13c2788ffe661f3101019ee63df86c0dbe1bfce76a Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.993190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-q6g28" event={"ID":"b1511901-0bd6-44e5-a199-c99065ee8138","Type":"ContainerStarted","Data":"23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320"} Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.993552 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-q6g28" event={"ID":"b1511901-0bd6-44e5-a199-c99065ee8138","Type":"ContainerStarted","Data":"cb8d0d0c67f1d975d677ac13c2788ffe661f3101019ee63df86c0dbe1bfce76a"} Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.995375 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-54852" Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.996021 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-54852" event={"ID":"55821f55-ec95-4d85-bc6d-c7aeef199f3c","Type":"ContainerDied","Data":"7cbeba730b5346ec21ea3bf94136dae208129f9d7c0c8e80dcd79cdab9d0d4c1"} Nov 27 07:24:25 crc kubenswrapper[4706]: I1127 07:24:25.996120 4706 scope.go:117] "RemoveContainer" containerID="b1e37dc1bac20e79e12376c3887c3fd4864cf2f01a71906618d02240b63cc3ce" Nov 27 07:24:26 crc kubenswrapper[4706]: I1127 07:24:26.003396 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" event={"ID":"525e7753-926d-429b-bbe0-f45e95cf80c3","Type":"ContainerStarted","Data":"71c31a5f98996cc2836816887d95d7c8ebb6f2af0069aead745b5894d3b87270"} Nov 27 07:24:26 crc kubenswrapper[4706]: I1127 07:24:26.014154 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-q6g28" podStartSLOduration=5.964449632 podStartE2EDuration="6.014130479s" podCreationTimestamp="2025-11-27 07:24:20 +0000 UTC" firstStartedPulling="2025-11-27 07:24:25.435500192 +0000 UTC m=+949.325091002" lastFinishedPulling="2025-11-27 07:24:25.485181039 +0000 UTC m=+949.374771849" observedRunningTime="2025-11-27 07:24:26.007035967 +0000 UTC m=+949.896626777" watchObservedRunningTime="2025-11-27 07:24:26.014130479 +0000 UTC m=+949.903721289" Nov 27 07:24:26 crc kubenswrapper[4706]: I1127 07:24:26.047655 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" podStartSLOduration=2.077895772 podStartE2EDuration="8.047634372s" podCreationTimestamp="2025-11-27 07:24:18 +0000 UTC" firstStartedPulling="2025-11-27 07:24:19.033113034 +0000 UTC m=+942.922703844" lastFinishedPulling="2025-11-27 07:24:25.002851634 +0000 UTC m=+948.892442444" observedRunningTime="2025-11-27 07:24:26.032558711 +0000 UTC m=+949.922149521" watchObservedRunningTime="2025-11-27 07:24:26.047634372 +0000 UTC m=+949.937225182" Nov 27 07:24:26 crc kubenswrapper[4706]: I1127 07:24:26.047812 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-54852"] Nov 27 07:24:26 crc kubenswrapper[4706]: I1127 07:24:26.055641 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-54852"] Nov 27 07:24:26 crc kubenswrapper[4706]: I1127 07:24:26.233655 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:26 crc kubenswrapper[4706]: E1127 07:24:26.233913 4706 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 27 07:24:26 crc kubenswrapper[4706]: E1127 07:24:26.233934 4706 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs: configmap "swift-ring-files" not found Nov 27 07:24:26 crc kubenswrapper[4706]: E1127 07:24:26.233992 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift podName:64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76 nodeName:}" failed. No retries permitted until 2025-11-27 07:24:34.233973429 +0000 UTC m=+958.123564239 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift") pod "swift-proxy-6bd58cfcf7-jd2hs" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76") : configmap "swift-ring-files" not found Nov 27 07:24:26 crc kubenswrapper[4706]: I1127 07:24:26.816821 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55821f55-ec95-4d85-bc6d-c7aeef199f3c" path="/var/lib/kubelet/pods/55821f55-ec95-4d85-bc6d-c7aeef199f3c/volumes" Nov 27 07:24:31 crc kubenswrapper[4706]: I1127 07:24:31.190692 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:31 crc kubenswrapper[4706]: I1127 07:24:31.191188 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:31 crc kubenswrapper[4706]: I1127 07:24:31.219771 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:32 crc kubenswrapper[4706]: I1127 07:24:32.060487 4706 generic.go:334] "Generic (PLEG): container finished" podID="525e7753-926d-429b-bbe0-f45e95cf80c3" containerID="71c31a5f98996cc2836816887d95d7c8ebb6f2af0069aead745b5894d3b87270" exitCode=0 Nov 27 07:24:32 crc kubenswrapper[4706]: I1127 07:24:32.060573 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" event={"ID":"525e7753-926d-429b-bbe0-f45e95cf80c3","Type":"ContainerDied","Data":"71c31a5f98996cc2836816887d95d7c8ebb6f2af0069aead745b5894d3b87270"} Nov 27 07:24:32 crc kubenswrapper[4706]: I1127 07:24:32.097888 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.307662 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.349444 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-dispersionconf\") pod \"525e7753-926d-429b-bbe0-f45e95cf80c3\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.349516 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/525e7753-926d-429b-bbe0-f45e95cf80c3-etc-swift\") pod \"525e7753-926d-429b-bbe0-f45e95cf80c3\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.349561 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5dgv\" (UniqueName: \"kubernetes.io/projected/525e7753-926d-429b-bbe0-f45e95cf80c3-kube-api-access-k5dgv\") pod \"525e7753-926d-429b-bbe0-f45e95cf80c3\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.349617 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-swiftconf\") pod \"525e7753-926d-429b-bbe0-f45e95cf80c3\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.349654 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-scripts\") pod \"525e7753-926d-429b-bbe0-f45e95cf80c3\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.349737 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-ring-data-devices\") pod \"525e7753-926d-429b-bbe0-f45e95cf80c3\" (UID: \"525e7753-926d-429b-bbe0-f45e95cf80c3\") " Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.350893 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "525e7753-926d-429b-bbe0-f45e95cf80c3" (UID: "525e7753-926d-429b-bbe0-f45e95cf80c3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.356136 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/525e7753-926d-429b-bbe0-f45e95cf80c3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "525e7753-926d-429b-bbe0-f45e95cf80c3" (UID: "525e7753-926d-429b-bbe0-f45e95cf80c3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.357366 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/525e7753-926d-429b-bbe0-f45e95cf80c3-kube-api-access-k5dgv" (OuterVolumeSpecName: "kube-api-access-k5dgv") pod "525e7753-926d-429b-bbe0-f45e95cf80c3" (UID: "525e7753-926d-429b-bbe0-f45e95cf80c3"). InnerVolumeSpecName "kube-api-access-k5dgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.371288 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-scripts" (OuterVolumeSpecName: "scripts") pod "525e7753-926d-429b-bbe0-f45e95cf80c3" (UID: "525e7753-926d-429b-bbe0-f45e95cf80c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.374921 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "525e7753-926d-429b-bbe0-f45e95cf80c3" (UID: "525e7753-926d-429b-bbe0-f45e95cf80c3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.375962 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "525e7753-926d-429b-bbe0-f45e95cf80c3" (UID: "525e7753-926d-429b-bbe0-f45e95cf80c3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.452129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.452328 4706 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.452343 4706 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/525e7753-926d-429b-bbe0-f45e95cf80c3-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.452356 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5dgv\" (UniqueName: \"kubernetes.io/projected/525e7753-926d-429b-bbe0-f45e95cf80c3-kube-api-access-k5dgv\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.452370 4706 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/525e7753-926d-429b-bbe0-f45e95cf80c3-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.452382 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.452393 4706 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/525e7753-926d-429b-bbe0-f45e95cf80c3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.456204 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"swift-storage-0\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.521999 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.903050 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd"] Nov 27 07:24:33 crc kubenswrapper[4706]: E1127 07:24:33.903676 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525e7753-926d-429b-bbe0-f45e95cf80c3" containerName="swift-ring-rebalance" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.903688 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="525e7753-926d-429b-bbe0-f45e95cf80c3" containerName="swift-ring-rebalance" Nov 27 07:24:33 crc kubenswrapper[4706]: E1127 07:24:33.903701 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55821f55-ec95-4d85-bc6d-c7aeef199f3c" containerName="registry-server" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.903707 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="55821f55-ec95-4d85-bc6d-c7aeef199f3c" containerName="registry-server" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.903821 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="55821f55-ec95-4d85-bc6d-c7aeef199f3c" containerName="registry-server" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.903834 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="525e7753-926d-429b-bbe0-f45e95cf80c3" containerName="swift-ring-rebalance" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.904689 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.910189 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dk9r5" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.916137 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd"] Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.957175 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.960115 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsgsq\" (UniqueName: \"kubernetes.io/projected/0917ef90-d580-44fd-b603-280fce1f2aec-kube-api-access-fsgsq\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.960179 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-util\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:33 crc kubenswrapper[4706]: I1127 07:24:33.960290 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-bundle\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.061779 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsgsq\" (UniqueName: \"kubernetes.io/projected/0917ef90-d580-44fd-b603-280fce1f2aec-kube-api-access-fsgsq\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.061833 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-util\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.061866 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-bundle\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.062624 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-bundle\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.062718 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-util\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.073603 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.073553 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-2zfbf" event={"ID":"525e7753-926d-429b-bbe0-f45e95cf80c3","Type":"ContainerDied","Data":"4a7c3b61ca649e490effdcfa590ba387faff7124c8b79bd0e2fed354229d0092"} Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.073783 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a7c3b61ca649e490effdcfa590ba387faff7124c8b79bd0e2fed354229d0092" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.074547 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"77b8887f2bd7f2407d49a3fd116c53505031d0d71c8ea57b6eddaeb41be8490f"} Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.077513 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsgsq\" (UniqueName: \"kubernetes.io/projected/0917ef90-d580-44fd-b603-280fce1f2aec-kube-api-access-fsgsq\") pod \"ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.228169 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.265723 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.270513 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jd2hs\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.277500 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.678555 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd"] Nov 27 07:24:34 crc kubenswrapper[4706]: I1127 07:24:34.731183 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs"] Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.082949 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" event={"ID":"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76","Type":"ContainerStarted","Data":"9f23d04861cc7026fe4bf45c52c3a91306ee4fea16da518473dfc8059c16ef9c"} Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.083244 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" event={"ID":"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76","Type":"ContainerStarted","Data":"fca707dfa183a719c29bd821db4190413d057073aa4d1266726fe8a84ee0fa4a"} Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.083915 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.083936 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.087137 4706 generic.go:334] "Generic (PLEG): container finished" podID="0917ef90-d580-44fd-b603-280fce1f2aec" containerID="3b4a2bd377e8eb5e0c81fe64c31c3c023dbb60de56a4ccc0dffa6544b8e63cb7" exitCode=0 Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.087170 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" event={"ID":"0917ef90-d580-44fd-b603-280fce1f2aec","Type":"ContainerDied","Data":"3b4a2bd377e8eb5e0c81fe64c31c3c023dbb60de56a4ccc0dffa6544b8e63cb7"} Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.087200 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" event={"ID":"0917ef90-d580-44fd-b603-280fce1f2aec","Type":"ContainerStarted","Data":"a95ac4895785c621aae64cf7aa0d4b6ef4b6da1b629d1463fc1ec8a8d0c6dc57"} Nov 27 07:24:35 crc kubenswrapper[4706]: I1127 07:24:35.121006 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" podStartSLOduration=17.120983536 podStartE2EDuration="17.120983536s" podCreationTimestamp="2025-11-27 07:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:24:35.099892065 +0000 UTC m=+958.989482875" watchObservedRunningTime="2025-11-27 07:24:35.120983536 +0000 UTC m=+959.010574356" Nov 27 07:24:36 crc kubenswrapper[4706]: I1127 07:24:36.110824 4706 generic.go:334] "Generic (PLEG): container finished" podID="0917ef90-d580-44fd-b603-280fce1f2aec" containerID="df440ed18630a9f558e95670a1d131f78d1180569d871106239545490b467850" exitCode=0 Nov 27 07:24:36 crc kubenswrapper[4706]: I1127 07:24:36.112348 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" event={"ID":"0917ef90-d580-44fd-b603-280fce1f2aec","Type":"ContainerDied","Data":"df440ed18630a9f558e95670a1d131f78d1180569d871106239545490b467850"} Nov 27 07:24:36 crc kubenswrapper[4706]: I1127 07:24:36.143204 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"b99dfd374a5ce1c13a934b2f873d212ff72dfe71b28341e2c047ff8d80e0c669"} Nov 27 07:24:36 crc kubenswrapper[4706]: I1127 07:24:36.143401 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"0854a4d87b476088ca363dd19e5424c00e3887a3cd94f9bb2324db18aa9630d0"} Nov 27 07:24:36 crc kubenswrapper[4706]: I1127 07:24:36.143505 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"be4e0386cd97f8aa803f6981bdc1c529197af63b34156666d8f8d7263d0d4f3d"} Nov 27 07:24:36 crc kubenswrapper[4706]: I1127 07:24:36.166518 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" event={"ID":"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76","Type":"ContainerStarted","Data":"c7619136346823904b27331aeb39977b3b1020d65267db3982003f2469a6f40c"} Nov 27 07:24:37 crc kubenswrapper[4706]: I1127 07:24:37.174106 4706 generic.go:334] "Generic (PLEG): container finished" podID="0917ef90-d580-44fd-b603-280fce1f2aec" containerID="2e9851ea626c3613fbaab116322ef135607b29f3b73e1c04c9b421e6edc66eab" exitCode=0 Nov 27 07:24:37 crc kubenswrapper[4706]: I1127 07:24:37.174181 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" event={"ID":"0917ef90-d580-44fd-b603-280fce1f2aec","Type":"ContainerDied","Data":"2e9851ea626c3613fbaab116322ef135607b29f3b73e1c04c9b421e6edc66eab"} Nov 27 07:24:37 crc kubenswrapper[4706]: I1127 07:24:37.178368 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"02d59e38675cdde34dc486df647b1eff69092a995a977d61d28b1c5da1ad6b1e"} Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.187348 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"e2ef352d28dc64cdc17f622c2a5ce4825d891748d60228a87b0234886b2fbd46"} Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.187658 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"de82cace4517a41df59b73d9a505b6d9bc00d26b88a1425eae7b033bb480c311"} Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.187672 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"de0a318ea00fadfa414f9b9cb4effedf4f8cbc78ffcf755bb0af2d1ee29bee87"} Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.187684 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"419c0ad19d57e0d824fc3545acc3c8fa4dc2b5da26669ee7c69100b12f6cf8ad"} Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.416238 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.443670 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-bundle\") pod \"0917ef90-d580-44fd-b603-280fce1f2aec\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.443773 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsgsq\" (UniqueName: \"kubernetes.io/projected/0917ef90-d580-44fd-b603-280fce1f2aec-kube-api-access-fsgsq\") pod \"0917ef90-d580-44fd-b603-280fce1f2aec\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.443852 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-util\") pod \"0917ef90-d580-44fd-b603-280fce1f2aec\" (UID: \"0917ef90-d580-44fd-b603-280fce1f2aec\") " Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.444904 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-bundle" (OuterVolumeSpecName: "bundle") pod "0917ef90-d580-44fd-b603-280fce1f2aec" (UID: "0917ef90-d580-44fd-b603-280fce1f2aec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.450245 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0917ef90-d580-44fd-b603-280fce1f2aec-kube-api-access-fsgsq" (OuterVolumeSpecName: "kube-api-access-fsgsq") pod "0917ef90-d580-44fd-b603-280fce1f2aec" (UID: "0917ef90-d580-44fd-b603-280fce1f2aec"). InnerVolumeSpecName "kube-api-access-fsgsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.459921 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-util" (OuterVolumeSpecName: "util") pod "0917ef90-d580-44fd-b603-280fce1f2aec" (UID: "0917ef90-d580-44fd-b603-280fce1f2aec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.545431 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsgsq\" (UniqueName: \"kubernetes.io/projected/0917ef90-d580-44fd-b603-280fce1f2aec-kube-api-access-fsgsq\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.545465 4706 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-util\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:38 crc kubenswrapper[4706]: I1127 07:24:38.545478 4706 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0917ef90-d580-44fd-b603-280fce1f2aec-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:24:39 crc kubenswrapper[4706]: I1127 07:24:39.200288 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" event={"ID":"0917ef90-d580-44fd-b603-280fce1f2aec","Type":"ContainerDied","Data":"a95ac4895785c621aae64cf7aa0d4b6ef4b6da1b629d1463fc1ec8a8d0c6dc57"} Nov 27 07:24:39 crc kubenswrapper[4706]: I1127 07:24:39.200542 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a95ac4895785c621aae64cf7aa0d4b6ef4b6da1b629d1463fc1ec8a8d0c6dc57" Nov 27 07:24:39 crc kubenswrapper[4706]: I1127 07:24:39.200340 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd" Nov 27 07:24:39 crc kubenswrapper[4706]: I1127 07:24:39.285110 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:40 crc kubenswrapper[4706]: I1127 07:24:40.221183 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"13bb352c512127bb19cb77c2407f696591efaeb94de88485c71ef25ef7efd32b"} Nov 27 07:24:40 crc kubenswrapper[4706]: I1127 07:24:40.221529 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"2078680d2bb75db1546bb97e0d090d85113a906cd7a2b21900140fb580300a81"} Nov 27 07:24:40 crc kubenswrapper[4706]: I1127 07:24:40.221539 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"b461410d2b38540e3971bd0c38fb4b2285b2b4bea2ad91957750db63eb8c977f"} Nov 27 07:24:40 crc kubenswrapper[4706]: I1127 07:24:40.221556 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"92f449c478541d4fc040be370d8fbfc4918ffb370dcd3a1fef9f2aeed113524a"} Nov 27 07:24:40 crc kubenswrapper[4706]: I1127 07:24:40.221567 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"600192eff2ee82990dc4e3bfc25d3f3cdde9d6e9447251423860a8d59ffdffb8"} Nov 27 07:24:41 crc kubenswrapper[4706]: I1127 07:24:41.235046 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"695ff04f9aad2b63580222fb53407fe88ab246ea589b941267e537a78643ddbd"} Nov 27 07:24:41 crc kubenswrapper[4706]: I1127 07:24:41.235448 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerStarted","Data":"87334b52ad9d36330c1c6887ab6aefaab2f4a4d21ec015aff098d9b7df9b66e5"} Nov 27 07:24:41 crc kubenswrapper[4706]: I1127 07:24:41.278183 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=20.035612238 podStartE2EDuration="25.278161955s" podCreationTimestamp="2025-11-27 07:24:16 +0000 UTC" firstStartedPulling="2025-11-27 07:24:33.966971339 +0000 UTC m=+957.856562149" lastFinishedPulling="2025-11-27 07:24:39.209521056 +0000 UTC m=+963.099111866" observedRunningTime="2025-11-27 07:24:41.270806926 +0000 UTC m=+965.160397776" watchObservedRunningTime="2025-11-27 07:24:41.278161955 +0000 UTC m=+965.167752765" Nov 27 07:24:44 crc kubenswrapper[4706]: I1127 07:24:44.281112 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.078953 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq"] Nov 27 07:24:56 crc kubenswrapper[4706]: E1127 07:24:56.079802 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" containerName="extract" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.079815 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" containerName="extract" Nov 27 07:24:56 crc kubenswrapper[4706]: E1127 07:24:56.079829 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" containerName="util" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.079834 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" containerName="util" Nov 27 07:24:56 crc kubenswrapper[4706]: E1127 07:24:56.079842 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" containerName="pull" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.079848 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" containerName="pull" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.079977 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" containerName="extract" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.080438 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.089742 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.089776 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mjj67" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.094288 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq"] Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.243976 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcvjt\" (UniqueName: \"kubernetes.io/projected/3b7ef303-eaeb-4081-a2b9-44fbbc439974-kube-api-access-tcvjt\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.244054 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-apiservice-cert\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.244095 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-webhook-cert\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.345347 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-apiservice-cert\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.345410 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-webhook-cert\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.345499 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcvjt\" (UniqueName: \"kubernetes.io/projected/3b7ef303-eaeb-4081-a2b9-44fbbc439974-kube-api-access-tcvjt\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.350494 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-apiservice-cert\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.350499 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-webhook-cert\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.364511 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcvjt\" (UniqueName: \"kubernetes.io/projected/3b7ef303-eaeb-4081-a2b9-44fbbc439974-kube-api-access-tcvjt\") pod \"glance-operator-controller-manager-84798567c5-8dsrq\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.417028 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:24:56 crc kubenswrapper[4706]: I1127 07:24:56.897281 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq"] Nov 27 07:24:57 crc kubenswrapper[4706]: I1127 07:24:57.353699 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" event={"ID":"3b7ef303-eaeb-4081-a2b9-44fbbc439974","Type":"ContainerStarted","Data":"bd89b2d90d72f608cf31d55f3feb1cd09de14170c1db4940ee648d2efb9a345e"} Nov 27 07:24:59 crc kubenswrapper[4706]: I1127 07:24:59.367965 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" event={"ID":"3b7ef303-eaeb-4081-a2b9-44fbbc439974","Type":"ContainerStarted","Data":"ed594cd76f55eb035d4feb6cb65c42fca83b863966f11dd5bcf5295a464b64bf"} Nov 27 07:24:59 crc kubenswrapper[4706]: I1127 07:24:59.368388 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:25:06 crc kubenswrapper[4706]: I1127 07:25:06.422510 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:25:06 crc kubenswrapper[4706]: I1127 07:25:06.443799 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" podStartSLOduration=9.113048264 podStartE2EDuration="10.44378088s" podCreationTimestamp="2025-11-27 07:24:56 +0000 UTC" firstStartedPulling="2025-11-27 07:24:56.902520145 +0000 UTC m=+980.792110965" lastFinishedPulling="2025-11-27 07:24:58.233252781 +0000 UTC m=+982.122843581" observedRunningTime="2025-11-27 07:24:59.39496479 +0000 UTC m=+983.284555600" watchObservedRunningTime="2025-11-27 07:25:06.44378088 +0000 UTC m=+990.333371700" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.519343 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-x9w4b"] Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.520656 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.525230 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd"] Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.526064 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.528062 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.535248 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd"] Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.550953 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.551819 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.561046 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.561334 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.561406 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-trwjt" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.569363 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.574651 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.583521 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-x9w4b"] Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.635878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed44d2c-d0dc-4c85-b9ff-810139ace09b-operator-scripts\") pod \"glance-db-create-x9w4b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.635937 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9b76\" (UniqueName: \"kubernetes.io/projected/aed44d2c-d0dc-4c85-b9ff-810139ace09b-kube-api-access-d9b76\") pod \"glance-db-create-x9w4b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.636276 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfxk6\" (UniqueName: \"kubernetes.io/projected/65045997-87c4-4ba0-b18a-9e656f61fc16-kube-api-access-gfxk6\") pod \"glance-b7e1-account-create-update-2pkrd\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.636337 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65045997-87c4-4ba0-b18a-9e656f61fc16-operator-scripts\") pod \"glance-b7e1-account-create-update-2pkrd\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.737706 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfxk6\" (UniqueName: \"kubernetes.io/projected/65045997-87c4-4ba0-b18a-9e656f61fc16-kube-api-access-gfxk6\") pod \"glance-b7e1-account-create-update-2pkrd\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.737770 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.737798 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqkdg\" (UniqueName: \"kubernetes.io/projected/4d54b307-a689-492f-99dd-9f7adb834498-kube-api-access-xqkdg\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.737902 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65045997-87c4-4ba0-b18a-9e656f61fc16-operator-scripts\") pod \"glance-b7e1-account-create-update-2pkrd\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.737948 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config-secret\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.737974 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed44d2c-d0dc-4c85-b9ff-810139ace09b-operator-scripts\") pod \"glance-db-create-x9w4b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.737992 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9b76\" (UniqueName: \"kubernetes.io/projected/aed44d2c-d0dc-4c85-b9ff-810139ace09b-kube-api-access-d9b76\") pod \"glance-db-create-x9w4b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.738020 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-scripts\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.738630 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65045997-87c4-4ba0-b18a-9e656f61fc16-operator-scripts\") pod \"glance-b7e1-account-create-update-2pkrd\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.738919 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed44d2c-d0dc-4c85-b9ff-810139ace09b-operator-scripts\") pod \"glance-db-create-x9w4b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.759862 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9b76\" (UniqueName: \"kubernetes.io/projected/aed44d2c-d0dc-4c85-b9ff-810139ace09b-kube-api-access-d9b76\") pod \"glance-db-create-x9w4b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.761635 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfxk6\" (UniqueName: \"kubernetes.io/projected/65045997-87c4-4ba0-b18a-9e656f61fc16-kube-api-access-gfxk6\") pod \"glance-b7e1-account-create-update-2pkrd\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.839124 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.839764 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config-secret\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.839886 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-scripts\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.839986 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.840079 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqkdg\" (UniqueName: \"kubernetes.io/projected/4d54b307-a689-492f-99dd-9f7adb834498-kube-api-access-xqkdg\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.841041 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.841649 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-scripts\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.846356 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.852740 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config-secret\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.864995 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqkdg\" (UniqueName: \"kubernetes.io/projected/4d54b307-a689-492f-99dd-9f7adb834498-kube-api-access-xqkdg\") pod \"openstackclient\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:09 crc kubenswrapper[4706]: I1127 07:25:09.882917 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.294900 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd"] Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.371367 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-x9w4b"] Nov 27 07:25:10 crc kubenswrapper[4706]: W1127 07:25:10.379364 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaed44d2c_d0dc_4c85_b9ff_810139ace09b.slice/crio-b375b861809ffc9f15e4072b556755ebb0662b51c52290586af653a423213c0c WatchSource:0}: Error finding container b375b861809ffc9f15e4072b556755ebb0662b51c52290586af653a423213c0c: Status 404 returned error can't find the container with id b375b861809ffc9f15e4072b556755ebb0662b51c52290586af653a423213c0c Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.421056 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:25:10 crc kubenswrapper[4706]: W1127 07:25:10.426502 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d54b307_a689_492f_99dd_9f7adb834498.slice/crio-b96679bf3c7f53cde29dabe6905449fa18fe344d94cbfc27260f978bd8c95c4d WatchSource:0}: Error finding container b96679bf3c7f53cde29dabe6905449fa18fe344d94cbfc27260f978bd8c95c4d: Status 404 returned error can't find the container with id b96679bf3c7f53cde29dabe6905449fa18fe344d94cbfc27260f978bd8c95c4d Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.448034 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" event={"ID":"65045997-87c4-4ba0-b18a-9e656f61fc16","Type":"ContainerStarted","Data":"c81e1bc161bc027db217be0c897cb083f59d3e9f287e621bd6d1eb3174b57ac3"} Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.448080 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" event={"ID":"65045997-87c4-4ba0-b18a-9e656f61fc16","Type":"ContainerStarted","Data":"5049e06288ceda5b2a132085a229d971c27806eb3741852d625c0c8f2779bf59"} Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.452680 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"4d54b307-a689-492f-99dd-9f7adb834498","Type":"ContainerStarted","Data":"b96679bf3c7f53cde29dabe6905449fa18fe344d94cbfc27260f978bd8c95c4d"} Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.454681 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-x9w4b" event={"ID":"aed44d2c-d0dc-4c85-b9ff-810139ace09b","Type":"ContainerStarted","Data":"b375b861809ffc9f15e4072b556755ebb0662b51c52290586af653a423213c0c"} Nov 27 07:25:10 crc kubenswrapper[4706]: I1127 07:25:10.468097 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" podStartSLOduration=1.468076937 podStartE2EDuration="1.468076937s" podCreationTimestamp="2025-11-27 07:25:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:25:10.460907374 +0000 UTC m=+994.350498184" watchObservedRunningTime="2025-11-27 07:25:10.468076937 +0000 UTC m=+994.357667747" Nov 27 07:25:11 crc kubenswrapper[4706]: I1127 07:25:11.464377 4706 generic.go:334] "Generic (PLEG): container finished" podID="65045997-87c4-4ba0-b18a-9e656f61fc16" containerID="c81e1bc161bc027db217be0c897cb083f59d3e9f287e621bd6d1eb3174b57ac3" exitCode=0 Nov 27 07:25:11 crc kubenswrapper[4706]: I1127 07:25:11.464417 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" event={"ID":"65045997-87c4-4ba0-b18a-9e656f61fc16","Type":"ContainerDied","Data":"c81e1bc161bc027db217be0c897cb083f59d3e9f287e621bd6d1eb3174b57ac3"} Nov 27 07:25:11 crc kubenswrapper[4706]: I1127 07:25:11.468010 4706 generic.go:334] "Generic (PLEG): container finished" podID="aed44d2c-d0dc-4c85-b9ff-810139ace09b" containerID="fa800bf58c6ed5a8d809f8f879a27d936f6231c341043933b237d54fc820e5ce" exitCode=0 Nov 27 07:25:11 crc kubenswrapper[4706]: I1127 07:25:11.468075 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-x9w4b" event={"ID":"aed44d2c-d0dc-4c85-b9ff-810139ace09b","Type":"ContainerDied","Data":"fa800bf58c6ed5a8d809f8f879a27d936f6231c341043933b237d54fc820e5ce"} Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.856033 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.860557 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.980463 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfxk6\" (UniqueName: \"kubernetes.io/projected/65045997-87c4-4ba0-b18a-9e656f61fc16-kube-api-access-gfxk6\") pod \"65045997-87c4-4ba0-b18a-9e656f61fc16\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.980555 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed44d2c-d0dc-4c85-b9ff-810139ace09b-operator-scripts\") pod \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.980584 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9b76\" (UniqueName: \"kubernetes.io/projected/aed44d2c-d0dc-4c85-b9ff-810139ace09b-kube-api-access-d9b76\") pod \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\" (UID: \"aed44d2c-d0dc-4c85-b9ff-810139ace09b\") " Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.980671 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65045997-87c4-4ba0-b18a-9e656f61fc16-operator-scripts\") pod \"65045997-87c4-4ba0-b18a-9e656f61fc16\" (UID: \"65045997-87c4-4ba0-b18a-9e656f61fc16\") " Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.981130 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65045997-87c4-4ba0-b18a-9e656f61fc16-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65045997-87c4-4ba0-b18a-9e656f61fc16" (UID: "65045997-87c4-4ba0-b18a-9e656f61fc16"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.981425 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed44d2c-d0dc-4c85-b9ff-810139ace09b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aed44d2c-d0dc-4c85-b9ff-810139ace09b" (UID: "aed44d2c-d0dc-4c85-b9ff-810139ace09b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.986183 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65045997-87c4-4ba0-b18a-9e656f61fc16-kube-api-access-gfxk6" (OuterVolumeSpecName: "kube-api-access-gfxk6") pod "65045997-87c4-4ba0-b18a-9e656f61fc16" (UID: "65045997-87c4-4ba0-b18a-9e656f61fc16"). InnerVolumeSpecName "kube-api-access-gfxk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:25:12 crc kubenswrapper[4706]: I1127 07:25:12.988132 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed44d2c-d0dc-4c85-b9ff-810139ace09b-kube-api-access-d9b76" (OuterVolumeSpecName: "kube-api-access-d9b76") pod "aed44d2c-d0dc-4c85-b9ff-810139ace09b" (UID: "aed44d2c-d0dc-4c85-b9ff-810139ace09b"). InnerVolumeSpecName "kube-api-access-d9b76". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.081866 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65045997-87c4-4ba0-b18a-9e656f61fc16-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.081897 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfxk6\" (UniqueName: \"kubernetes.io/projected/65045997-87c4-4ba0-b18a-9e656f61fc16-kube-api-access-gfxk6\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.081908 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed44d2c-d0dc-4c85-b9ff-810139ace09b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.081918 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9b76\" (UniqueName: \"kubernetes.io/projected/aed44d2c-d0dc-4c85-b9ff-810139ace09b-kube-api-access-d9b76\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.486773 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-x9w4b" event={"ID":"aed44d2c-d0dc-4c85-b9ff-810139ace09b","Type":"ContainerDied","Data":"b375b861809ffc9f15e4072b556755ebb0662b51c52290586af653a423213c0c"} Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.486814 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b375b861809ffc9f15e4072b556755ebb0662b51c52290586af653a423213c0c" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.486864 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x9w4b" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.494026 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" event={"ID":"65045997-87c4-4ba0-b18a-9e656f61fc16","Type":"ContainerDied","Data":"5049e06288ceda5b2a132085a229d971c27806eb3741852d625c0c8f2779bf59"} Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.494068 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5049e06288ceda5b2a132085a229d971c27806eb3741852d625c0c8f2779bf59" Nov 27 07:25:13 crc kubenswrapper[4706]: I1127 07:25:13.494072 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.646277 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-rgdxk"] Nov 27 07:25:14 crc kubenswrapper[4706]: E1127 07:25:14.646606 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed44d2c-d0dc-4c85-b9ff-810139ace09b" containerName="mariadb-database-create" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.646618 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed44d2c-d0dc-4c85-b9ff-810139ace09b" containerName="mariadb-database-create" Nov 27 07:25:14 crc kubenswrapper[4706]: E1127 07:25:14.646632 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65045997-87c4-4ba0-b18a-9e656f61fc16" containerName="mariadb-account-create-update" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.646638 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="65045997-87c4-4ba0-b18a-9e656f61fc16" containerName="mariadb-account-create-update" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.646750 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed44d2c-d0dc-4c85-b9ff-810139ace09b" containerName="mariadb-database-create" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.646766 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="65045997-87c4-4ba0-b18a-9e656f61fc16" containerName="mariadb-account-create-update" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.647192 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.650153 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.656666 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-g9p2j" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.673647 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rgdxk"] Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.818838 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-db-sync-config-data\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.819132 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfmnh\" (UniqueName: \"kubernetes.io/projected/bccd6970-f332-4b8b-841c-a328277607ce-kube-api-access-bfmnh\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.819311 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-config-data\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.921043 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-db-sync-config-data\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.921386 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfmnh\" (UniqueName: \"kubernetes.io/projected/bccd6970-f332-4b8b-841c-a328277607ce-kube-api-access-bfmnh\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.921500 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-config-data\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.927015 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-db-sync-config-data\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.937630 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-config-data\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.937967 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfmnh\" (UniqueName: \"kubernetes.io/projected/bccd6970-f332-4b8b-841c-a328277607ce-kube-api-access-bfmnh\") pod \"glance-db-sync-rgdxk\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:14 crc kubenswrapper[4706]: I1127 07:25:14.974044 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:18 crc kubenswrapper[4706]: I1127 07:25:18.873254 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rgdxk"] Nov 27 07:25:19 crc kubenswrapper[4706]: I1127 07:25:19.546624 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rgdxk" event={"ID":"bccd6970-f332-4b8b-841c-a328277607ce","Type":"ContainerStarted","Data":"7627e65eb1b6b40014797b3ff998a8c3e55d7725ef08658d8750bf036ff32261"} Nov 27 07:25:19 crc kubenswrapper[4706]: I1127 07:25:19.548834 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"4d54b307-a689-492f-99dd-9f7adb834498","Type":"ContainerStarted","Data":"69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12"} Nov 27 07:25:19 crc kubenswrapper[4706]: I1127 07:25:19.581634 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.520171547 podStartE2EDuration="10.581605563s" podCreationTimestamp="2025-11-27 07:25:09 +0000 UTC" firstStartedPulling="2025-11-27 07:25:10.428351028 +0000 UTC m=+994.317941838" lastFinishedPulling="2025-11-27 07:25:18.489785044 +0000 UTC m=+1002.379375854" observedRunningTime="2025-11-27 07:25:19.573999086 +0000 UTC m=+1003.463589926" watchObservedRunningTime="2025-11-27 07:25:19.581605563 +0000 UTC m=+1003.471196373" Nov 27 07:25:30 crc kubenswrapper[4706]: I1127 07:25:30.633165 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rgdxk" event={"ID":"bccd6970-f332-4b8b-841c-a328277607ce","Type":"ContainerStarted","Data":"c7f09509e6118431d8aa40dd5faf8bb79c5b34059e6435631f0f60e1c386fc32"} Nov 27 07:25:30 crc kubenswrapper[4706]: I1127 07:25:30.662952 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-rgdxk" podStartSLOduration=6.155157405 podStartE2EDuration="16.662921928s" podCreationTimestamp="2025-11-27 07:25:14 +0000 UTC" firstStartedPulling="2025-11-27 07:25:18.876527166 +0000 UTC m=+1002.766117976" lastFinishedPulling="2025-11-27 07:25:29.384291689 +0000 UTC m=+1013.273882499" observedRunningTime="2025-11-27 07:25:30.658152689 +0000 UTC m=+1014.547743499" watchObservedRunningTime="2025-11-27 07:25:30.662921928 +0000 UTC m=+1014.552512728" Nov 27 07:25:38 crc kubenswrapper[4706]: I1127 07:25:38.707717 4706 generic.go:334] "Generic (PLEG): container finished" podID="bccd6970-f332-4b8b-841c-a328277607ce" containerID="c7f09509e6118431d8aa40dd5faf8bb79c5b34059e6435631f0f60e1c386fc32" exitCode=0 Nov 27 07:25:38 crc kubenswrapper[4706]: I1127 07:25:38.707808 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rgdxk" event={"ID":"bccd6970-f332-4b8b-841c-a328277607ce","Type":"ContainerDied","Data":"c7f09509e6118431d8aa40dd5faf8bb79c5b34059e6435631f0f60e1c386fc32"} Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.042260 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.126148 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-config-data\") pod \"bccd6970-f332-4b8b-841c-a328277607ce\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.126261 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfmnh\" (UniqueName: \"kubernetes.io/projected/bccd6970-f332-4b8b-841c-a328277607ce-kube-api-access-bfmnh\") pod \"bccd6970-f332-4b8b-841c-a328277607ce\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.126305 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-db-sync-config-data\") pod \"bccd6970-f332-4b8b-841c-a328277607ce\" (UID: \"bccd6970-f332-4b8b-841c-a328277607ce\") " Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.130732 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bccd6970-f332-4b8b-841c-a328277607ce-kube-api-access-bfmnh" (OuterVolumeSpecName: "kube-api-access-bfmnh") pod "bccd6970-f332-4b8b-841c-a328277607ce" (UID: "bccd6970-f332-4b8b-841c-a328277607ce"). InnerVolumeSpecName "kube-api-access-bfmnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.131353 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bccd6970-f332-4b8b-841c-a328277607ce" (UID: "bccd6970-f332-4b8b-841c-a328277607ce"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.162205 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-config-data" (OuterVolumeSpecName: "config-data") pod "bccd6970-f332-4b8b-841c-a328277607ce" (UID: "bccd6970-f332-4b8b-841c-a328277607ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.228132 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.228166 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccd6970-f332-4b8b-841c-a328277607ce-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.228180 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfmnh\" (UniqueName: \"kubernetes.io/projected/bccd6970-f332-4b8b-841c-a328277607ce-kube-api-access-bfmnh\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.726829 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rgdxk" event={"ID":"bccd6970-f332-4b8b-841c-a328277607ce","Type":"ContainerDied","Data":"7627e65eb1b6b40014797b3ff998a8c3e55d7725ef08658d8750bf036ff32261"} Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.726868 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7627e65eb1b6b40014797b3ff998a8c3e55d7725ef08658d8750bf036ff32261" Nov 27 07:25:40 crc kubenswrapper[4706]: I1127 07:25:40.726937 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rgdxk" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.970488 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:25:41 crc kubenswrapper[4706]: E1127 07:25:41.971110 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bccd6970-f332-4b8b-841c-a328277607ce" containerName="glance-db-sync" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.971122 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bccd6970-f332-4b8b-841c-a328277607ce" containerName="glance-db-sync" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.971264 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bccd6970-f332-4b8b-841c-a328277607ce" containerName="glance-db-sync" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.971909 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.973781 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-g9p2j" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.973922 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.981567 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 27 07:25:41 crc kubenswrapper[4706]: I1127 07:25:41.985212 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056285 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-dev\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056338 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056364 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-config-data\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056393 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-lib-modules\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056419 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-httpd-run\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056616 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056638 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056658 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-run\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056674 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-scripts\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056689 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056710 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-sys\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056724 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-nvme\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056737 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-logs\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.056755 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6xnk\" (UniqueName: \"kubernetes.io/projected/bd814f44-0460-400d-8570-594de21b5846-kube-api-access-j6xnk\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.108676 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.109948 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.122367 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158349 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-lib-modules\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158444 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-httpd-run\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158499 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-lib-modules\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158630 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-run\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158671 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-scripts\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158716 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158793 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-sys\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-nvme\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158844 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-logs\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.158878 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6xnk\" (UniqueName: \"kubernetes.io/projected/bd814f44-0460-400d-8570-594de21b5846-kube-api-access-j6xnk\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159019 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-httpd-run\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159027 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159078 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159112 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-dev\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159241 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159282 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-config-data\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159675 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159840 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-nvme\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159111 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-sys\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.159969 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-run\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.160010 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-dev\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.160153 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.162997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-logs\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.170809 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-scripts\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.171138 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-config-data\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.175670 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6xnk\" (UniqueName: \"kubernetes.io/projected/bd814f44-0460-400d-8570-594de21b5846-kube-api-access-j6xnk\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.186281 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.187498 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261105 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-scripts\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261151 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261173 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261189 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-httpd-run\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261480 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-config-data\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261585 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-run\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261711 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-dev\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261766 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261837 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261856 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-sys\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.261967 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-logs\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.262033 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-lib-modules\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.262196 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-nvme\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.262245 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-498t6\" (UniqueName: \"kubernetes.io/projected/35d07a7e-22b4-422f-bdf1-5b21a244ca44-kube-api-access-498t6\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.310426 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364114 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-498t6\" (UniqueName: \"kubernetes.io/projected/35d07a7e-22b4-422f-bdf1-5b21a244ca44-kube-api-access-498t6\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364169 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-nvme\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364205 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-scripts\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364244 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364273 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364298 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-httpd-run\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364337 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-config-data\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364358 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-run\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364395 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-dev\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364416 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364439 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364446 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364458 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-sys\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364494 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-logs\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364524 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-lib-modules\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364559 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364519 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-nvme\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.364713 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.365013 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-httpd-run\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.365091 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-logs\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.365108 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.365149 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-sys\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.365151 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-run\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.365177 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-lib-modules\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.365191 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-dev\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.370443 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-config-data\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.373027 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-scripts\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.383934 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.385894 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-498t6\" (UniqueName: \"kubernetes.io/projected/35d07a7e-22b4-422f-bdf1-5b21a244ca44-kube-api-access-498t6\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.387459 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.426248 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.656483 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:42 crc kubenswrapper[4706]: W1127 07:25:42.661533 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35d07a7e_22b4_422f_bdf1_5b21a244ca44.slice/crio-7e04c5a5eb0c2bab55a2ee59b81bd6b0be21f3a99318b6b264aa8ed43755e184 WatchSource:0}: Error finding container 7e04c5a5eb0c2bab55a2ee59b81bd6b0be21f3a99318b6b264aa8ed43755e184: Status 404 returned error can't find the container with id 7e04c5a5eb0c2bab55a2ee59b81bd6b0be21f3a99318b6b264aa8ed43755e184 Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.750897 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"35d07a7e-22b4-422f-bdf1-5b21a244ca44","Type":"ContainerStarted","Data":"7e04c5a5eb0c2bab55a2ee59b81bd6b0be21f3a99318b6b264aa8ed43755e184"} Nov 27 07:25:42 crc kubenswrapper[4706]: I1127 07:25:42.761356 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:25:42 crc kubenswrapper[4706]: W1127 07:25:42.764256 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd814f44_0460_400d_8570_594de21b5846.slice/crio-06269cef2d4a387255f943064d9734d6c1d8b1cd7d4eb59f1de134f94a61c081 WatchSource:0}: Error finding container 06269cef2d4a387255f943064d9734d6c1d8b1cd7d4eb59f1de134f94a61c081: Status 404 returned error can't find the container with id 06269cef2d4a387255f943064d9734d6c1d8b1cd7d4eb59f1de134f94a61c081 Nov 27 07:25:43 crc kubenswrapper[4706]: I1127 07:25:43.760647 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"35d07a7e-22b4-422f-bdf1-5b21a244ca44","Type":"ContainerStarted","Data":"67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea"} Nov 27 07:25:43 crc kubenswrapper[4706]: I1127 07:25:43.761417 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"35d07a7e-22b4-422f-bdf1-5b21a244ca44","Type":"ContainerStarted","Data":"42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3"} Nov 27 07:25:43 crc kubenswrapper[4706]: I1127 07:25:43.762454 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"bd814f44-0460-400d-8570-594de21b5846","Type":"ContainerStarted","Data":"fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280"} Nov 27 07:25:43 crc kubenswrapper[4706]: I1127 07:25:43.762507 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"bd814f44-0460-400d-8570-594de21b5846","Type":"ContainerStarted","Data":"b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651"} Nov 27 07:25:43 crc kubenswrapper[4706]: I1127 07:25:43.762523 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"bd814f44-0460-400d-8570-594de21b5846","Type":"ContainerStarted","Data":"06269cef2d4a387255f943064d9734d6c1d8b1cd7d4eb59f1de134f94a61c081"} Nov 27 07:25:43 crc kubenswrapper[4706]: I1127 07:25:43.798201 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.79818183 podStartE2EDuration="2.79818183s" podCreationTimestamp="2025-11-27 07:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:25:43.786595948 +0000 UTC m=+1027.676186778" watchObservedRunningTime="2025-11-27 07:25:43.79818183 +0000 UTC m=+1027.687772650" Nov 27 07:25:43 crc kubenswrapper[4706]: I1127 07:25:43.815404 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.815386927 podStartE2EDuration="2.815386927s" podCreationTimestamp="2025-11-27 07:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:25:43.81421275 +0000 UTC m=+1027.703803560" watchObservedRunningTime="2025-11-27 07:25:43.815386927 +0000 UTC m=+1027.704977737" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.310895 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.311422 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.338881 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.364852 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.427615 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.427659 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.470006 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.484998 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.825485 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.825718 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.826021 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:52 crc kubenswrapper[4706]: I1127 07:25:52.826096 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.801817 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.836612 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.883108 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.933911 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.934052 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.934065 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.934211 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-log" containerID="cri-o://42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3" gracePeriod=30 Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.934704 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-httpd" containerID="cri-o://67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea" gracePeriod=30 Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.976925 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": EOF" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.986589 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": EOF" Nov 27 07:25:54 crc kubenswrapper[4706]: I1127 07:25:54.986589 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": EOF" Nov 27 07:25:55 crc kubenswrapper[4706]: I1127 07:25:55.846342 4706 generic.go:334] "Generic (PLEG): container finished" podID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerID="42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3" exitCode=143 Nov 27 07:25:55 crc kubenswrapper[4706]: I1127 07:25:55.846410 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"35d07a7e-22b4-422f-bdf1-5b21a244ca44","Type":"ContainerDied","Data":"42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3"} Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.092869 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": read tcp 10.217.0.2:48408->10.217.0.101:9292: read: connection reset by peer" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.422313 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573608 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-498t6\" (UniqueName: \"kubernetes.io/projected/35d07a7e-22b4-422f-bdf1-5b21a244ca44-kube-api-access-498t6\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573683 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-iscsi\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573708 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-nvme\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573738 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-scripts\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573770 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-httpd-run\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573788 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-run\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573803 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-lib-modules\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573823 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-logs\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573850 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-dev\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573875 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-sys\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573891 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573911 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573927 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-var-locks-brick\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.573951 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-config-data\") pod \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\" (UID: \"35d07a7e-22b4-422f-bdf1-5b21a244ca44\") " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.574338 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.574394 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.574414 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.574857 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-sys" (OuterVolumeSpecName: "sys") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.574929 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.575060 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-dev" (OuterVolumeSpecName: "dev") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.575089 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-run" (OuterVolumeSpecName: "run") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.575109 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.575399 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-logs" (OuterVolumeSpecName: "logs") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.579350 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-scripts" (OuterVolumeSpecName: "scripts") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.579845 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.589382 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.590562 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d07a7e-22b4-422f-bdf1-5b21a244ca44-kube-api-access-498t6" (OuterVolumeSpecName: "kube-api-access-498t6") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "kube-api-access-498t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.626757 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-config-data" (OuterVolumeSpecName: "config-data") pod "35d07a7e-22b4-422f-bdf1-5b21a244ca44" (UID: "35d07a7e-22b4-422f-bdf1-5b21a244ca44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675334 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675369 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675401 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675416 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675426 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675435 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675444 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-498t6\" (UniqueName: \"kubernetes.io/projected/35d07a7e-22b4-422f-bdf1-5b21a244ca44-kube-api-access-498t6\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675453 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675462 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675471 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d07a7e-22b4-422f-bdf1-5b21a244ca44-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675479 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675486 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675493 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35d07a7e-22b4-422f-bdf1-5b21a244ca44-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.675500 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35d07a7e-22b4-422f-bdf1-5b21a244ca44-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.689186 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.692730 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.777367 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.777402 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.867807 4706 generic.go:334] "Generic (PLEG): container finished" podID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerID="67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea" exitCode=0 Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.867852 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"35d07a7e-22b4-422f-bdf1-5b21a244ca44","Type":"ContainerDied","Data":"67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea"} Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.867878 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"35d07a7e-22b4-422f-bdf1-5b21a244ca44","Type":"ContainerDied","Data":"7e04c5a5eb0c2bab55a2ee59b81bd6b0be21f3a99318b6b264aa8ed43755e184"} Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.867876 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.867961 4706 scope.go:117] "RemoveContainer" containerID="67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.895200 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.900912 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.914467 4706 scope.go:117] "RemoveContainer" containerID="42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.936022 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:58 crc kubenswrapper[4706]: E1127 07:25:58.936405 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-log" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.936421 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-log" Nov 27 07:25:58 crc kubenswrapper[4706]: E1127 07:25:58.936433 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-httpd" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.936441 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-httpd" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.936622 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-log" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.936640 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" containerName="glance-httpd" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.937478 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.946140 4706 scope.go:117] "RemoveContainer" containerID="67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea" Nov 27 07:25:58 crc kubenswrapper[4706]: E1127 07:25:58.946588 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea\": container with ID starting with 67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea not found: ID does not exist" containerID="67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.946626 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea"} err="failed to get container status \"67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea\": rpc error: code = NotFound desc = could not find container \"67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea\": container with ID starting with 67c1c36a685a2e7273863fb17c65338e776e10b88fab8fde390448ba6260efea not found: ID does not exist" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.946650 4706 scope.go:117] "RemoveContainer" containerID="42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3" Nov 27 07:25:58 crc kubenswrapper[4706]: E1127 07:25:58.946940 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3\": container with ID starting with 42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3 not found: ID does not exist" containerID="42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.946972 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3"} err="failed to get container status \"42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3\": rpc error: code = NotFound desc = could not find container \"42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3\": container with ID starting with 42e9911b9e6aad7feb5a83fa7a6f150bd21642a81acbcfff62e028551f7539d3 not found: ID does not exist" Nov 27 07:25:58 crc kubenswrapper[4706]: I1127 07:25:58.965326 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081479 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-dev\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081518 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-config-data\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081567 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-sys\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081583 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081599 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081644 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081675 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-scripts\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081689 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-run\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081703 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-logs\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081721 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cc4g\" (UniqueName: \"kubernetes.io/projected/18672d3a-3a06-4450-bad8-2048f1afc942-kube-api-access-5cc4g\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081757 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-nvme\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081771 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-lib-modules\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081790 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-httpd-run\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.081862 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183242 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183310 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-dev\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183331 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-config-data\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183378 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-sys\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183397 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183417 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183434 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183468 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-scripts\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183482 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-dev\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183539 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-run\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183560 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183597 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183608 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183505 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-sys\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183596 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183494 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-run\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183759 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-logs\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183796 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cc4g\" (UniqueName: \"kubernetes.io/projected/18672d3a-3a06-4450-bad8-2048f1afc942-kube-api-access-5cc4g\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183835 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-nvme\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183858 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-lib-modules\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.183881 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-httpd-run\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.184189 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-logs\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.184252 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-nvme\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.184510 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-lib-modules\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.184723 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-httpd-run\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.190164 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-scripts\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.192614 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-config-data\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.203400 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.213835 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cc4g\" (UniqueName: \"kubernetes.io/projected/18672d3a-3a06-4450-bad8-2048f1afc942-kube-api-access-5cc4g\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.214850 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.263902 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.741458 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:25:59 crc kubenswrapper[4706]: W1127 07:25:59.750450 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18672d3a_3a06_4450_bad8_2048f1afc942.slice/crio-6f50833712b942cf4a28966f15b3dea5d4d044ccc414cbe86c89403eafe3497c WatchSource:0}: Error finding container 6f50833712b942cf4a28966f15b3dea5d4d044ccc414cbe86c89403eafe3497c: Status 404 returned error can't find the container with id 6f50833712b942cf4a28966f15b3dea5d4d044ccc414cbe86c89403eafe3497c Nov 27 07:25:59 crc kubenswrapper[4706]: I1127 07:25:59.876420 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"18672d3a-3a06-4450-bad8-2048f1afc942","Type":"ContainerStarted","Data":"6f50833712b942cf4a28966f15b3dea5d4d044ccc414cbe86c89403eafe3497c"} Nov 27 07:26:00 crc kubenswrapper[4706]: I1127 07:26:00.786830 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d07a7e-22b4-422f-bdf1-5b21a244ca44" path="/var/lib/kubelet/pods/35d07a7e-22b4-422f-bdf1-5b21a244ca44/volumes" Nov 27 07:26:00 crc kubenswrapper[4706]: I1127 07:26:00.886965 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"18672d3a-3a06-4450-bad8-2048f1afc942","Type":"ContainerStarted","Data":"21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05"} Nov 27 07:26:00 crc kubenswrapper[4706]: I1127 07:26:00.887268 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"18672d3a-3a06-4450-bad8-2048f1afc942","Type":"ContainerStarted","Data":"5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584"} Nov 27 07:26:00 crc kubenswrapper[4706]: I1127 07:26:00.925000 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.924977727 podStartE2EDuration="2.924977727s" podCreationTimestamp="2025-11-27 07:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:26:00.915993505 +0000 UTC m=+1044.805584355" watchObservedRunningTime="2025-11-27 07:26:00.924977727 +0000 UTC m=+1044.814568547" Nov 27 07:26:09 crc kubenswrapper[4706]: I1127 07:26:09.265170 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:09 crc kubenswrapper[4706]: I1127 07:26:09.266190 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:09 crc kubenswrapper[4706]: I1127 07:26:09.316369 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:09 crc kubenswrapper[4706]: I1127 07:26:09.339248 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:09 crc kubenswrapper[4706]: I1127 07:26:09.980843 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:09 crc kubenswrapper[4706]: I1127 07:26:09.980940 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:11 crc kubenswrapper[4706]: I1127 07:26:11.872841 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:11 crc kubenswrapper[4706]: I1127 07:26:11.873709 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:15 crc kubenswrapper[4706]: I1127 07:26:15.178332 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:26:15 crc kubenswrapper[4706]: I1127 07:26:15.178811 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:26:39 crc kubenswrapper[4706]: I1127 07:26:39.993350 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rgdxk"] Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.002826 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rgdxk"] Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.075533 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.075930 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-httpd" containerID="cri-o://fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280" gracePeriod=30 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.075943 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-log" containerID="cri-o://b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651" gracePeriod=30 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.085495 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.085840 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-log" containerID="cri-o://5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584" gracePeriod=30 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.085947 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-httpd" containerID="cri-o://21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05" gracePeriod=30 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.106736 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanceb7e1-account-delete-7crsq"] Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.107663 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.120804 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceb7e1-account-delete-7crsq"] Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.135787 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea7c58b-3f82-4e10-9e02-12fde72999ca-operator-scripts\") pod \"glanceb7e1-account-delete-7crsq\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.135862 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shl5m\" (UniqueName: \"kubernetes.io/projected/1ea7c58b-3f82-4e10-9e02-12fde72999ca-kube-api-access-shl5m\") pod \"glanceb7e1-account-delete-7crsq\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.218003 4706 generic.go:334] "Generic (PLEG): container finished" podID="bd814f44-0460-400d-8570-594de21b5846" containerID="b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651" exitCode=143 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.218308 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"bd814f44-0460-400d-8570-594de21b5846","Type":"ContainerDied","Data":"b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651"} Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.230091 4706 generic.go:334] "Generic (PLEG): container finished" podID="18672d3a-3a06-4450-bad8-2048f1afc942" containerID="5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584" exitCode=143 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.230137 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"18672d3a-3a06-4450-bad8-2048f1afc942","Type":"ContainerDied","Data":"5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584"} Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.236521 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shl5m\" (UniqueName: \"kubernetes.io/projected/1ea7c58b-3f82-4e10-9e02-12fde72999ca-kube-api-access-shl5m\") pod \"glanceb7e1-account-delete-7crsq\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.236614 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea7c58b-3f82-4e10-9e02-12fde72999ca-operator-scripts\") pod \"glanceb7e1-account-delete-7crsq\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.237414 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea7c58b-3f82-4e10-9e02-12fde72999ca-operator-scripts\") pod \"glanceb7e1-account-delete-7crsq\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.239957 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.240159 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="4d54b307-a689-492f-99dd-9f7adb834498" containerName="openstackclient" containerID="cri-o://69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12" gracePeriod=30 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.255962 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shl5m\" (UniqueName: \"kubernetes.io/projected/1ea7c58b-3f82-4e10-9e02-12fde72999ca-kube-api-access-shl5m\") pod \"glanceb7e1-account-delete-7crsq\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.429259 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.568665 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.664792 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceb7e1-account-delete-7crsq"] Nov 27 07:26:40 crc kubenswrapper[4706]: W1127 07:26:40.670087 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ea7c58b_3f82_4e10_9e02_12fde72999ca.slice/crio-4eeef396afeb135bb15068f6cecc9085389e8f33668f51dbd47d65d4aa41e341 WatchSource:0}: Error finding container 4eeef396afeb135bb15068f6cecc9085389e8f33668f51dbd47d65d4aa41e341: Status 404 returned error can't find the container with id 4eeef396afeb135bb15068f6cecc9085389e8f33668f51dbd47d65d4aa41e341 Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.743195 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqkdg\" (UniqueName: \"kubernetes.io/projected/4d54b307-a689-492f-99dd-9f7adb834498-kube-api-access-xqkdg\") pod \"4d54b307-a689-492f-99dd-9f7adb834498\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.743309 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config\") pod \"4d54b307-a689-492f-99dd-9f7adb834498\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.743331 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-scripts\") pod \"4d54b307-a689-492f-99dd-9f7adb834498\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.743427 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config-secret\") pod \"4d54b307-a689-492f-99dd-9f7adb834498\" (UID: \"4d54b307-a689-492f-99dd-9f7adb834498\") " Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.744158 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "4d54b307-a689-492f-99dd-9f7adb834498" (UID: "4d54b307-a689-492f-99dd-9f7adb834498"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.748702 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d54b307-a689-492f-99dd-9f7adb834498-kube-api-access-xqkdg" (OuterVolumeSpecName: "kube-api-access-xqkdg") pod "4d54b307-a689-492f-99dd-9f7adb834498" (UID: "4d54b307-a689-492f-99dd-9f7adb834498"). InnerVolumeSpecName "kube-api-access-xqkdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.764253 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4d54b307-a689-492f-99dd-9f7adb834498" (UID: "4d54b307-a689-492f-99dd-9f7adb834498"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.765213 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4d54b307-a689-492f-99dd-9f7adb834498" (UID: "4d54b307-a689-492f-99dd-9f7adb834498"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.784846 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bccd6970-f332-4b8b-841c-a328277607ce" path="/var/lib/kubelet/pods/bccd6970-f332-4b8b-841c-a328277607ce/volumes" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.845505 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.845542 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqkdg\" (UniqueName: \"kubernetes.io/projected/4d54b307-a689-492f-99dd-9f7adb834498-kube-api-access-xqkdg\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.845557 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:40 crc kubenswrapper[4706]: I1127 07:26:40.845570 4706 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d54b307-a689-492f-99dd-9f7adb834498-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.237742 4706 generic.go:334] "Generic (PLEG): container finished" podID="4d54b307-a689-492f-99dd-9f7adb834498" containerID="69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12" exitCode=143 Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.237853 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.237848 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"4d54b307-a689-492f-99dd-9f7adb834498","Type":"ContainerDied","Data":"69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12"} Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.238176 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"4d54b307-a689-492f-99dd-9f7adb834498","Type":"ContainerDied","Data":"b96679bf3c7f53cde29dabe6905449fa18fe344d94cbfc27260f978bd8c95c4d"} Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.238196 4706 scope.go:117] "RemoveContainer" containerID="69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12" Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.240314 4706 generic.go:334] "Generic (PLEG): container finished" podID="1ea7c58b-3f82-4e10-9e02-12fde72999ca" containerID="b13b57ab0e5f398cb514f76ffe3b76d52d898f85bb3464ae1c021ac4f13b27a2" exitCode=0 Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.240341 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" event={"ID":"1ea7c58b-3f82-4e10-9e02-12fde72999ca","Type":"ContainerDied","Data":"b13b57ab0e5f398cb514f76ffe3b76d52d898f85bb3464ae1c021ac4f13b27a2"} Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.240372 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" event={"ID":"1ea7c58b-3f82-4e10-9e02-12fde72999ca","Type":"ContainerStarted","Data":"4eeef396afeb135bb15068f6cecc9085389e8f33668f51dbd47d65d4aa41e341"} Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.261385 4706 scope.go:117] "RemoveContainer" containerID="69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12" Nov 27 07:26:41 crc kubenswrapper[4706]: E1127 07:26:41.261914 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12\": container with ID starting with 69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12 not found: ID does not exist" containerID="69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12" Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.261973 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12"} err="failed to get container status \"69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12\": rpc error: code = NotFound desc = could not find container \"69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12\": container with ID starting with 69c90d533df32a92fabba0e0b5b2e607b4f3e020520d178d5116eeab26512c12 not found: ID does not exist" Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.273108 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:26:41 crc kubenswrapper[4706]: I1127 07:26:41.277897 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.658018 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.776437 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea7c58b-3f82-4e10-9e02-12fde72999ca-operator-scripts\") pod \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.776640 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shl5m\" (UniqueName: \"kubernetes.io/projected/1ea7c58b-3f82-4e10-9e02-12fde72999ca-kube-api-access-shl5m\") pod \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\" (UID: \"1ea7c58b-3f82-4e10-9e02-12fde72999ca\") " Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.777745 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea7c58b-3f82-4e10-9e02-12fde72999ca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ea7c58b-3f82-4e10-9e02-12fde72999ca" (UID: "1ea7c58b-3f82-4e10-9e02-12fde72999ca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.782529 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea7c58b-3f82-4e10-9e02-12fde72999ca-kube-api-access-shl5m" (OuterVolumeSpecName: "kube-api-access-shl5m") pod "1ea7c58b-3f82-4e10-9e02-12fde72999ca" (UID: "1ea7c58b-3f82-4e10-9e02-12fde72999ca"). InnerVolumeSpecName "kube-api-access-shl5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.786409 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d54b307-a689-492f-99dd-9f7adb834498" path="/var/lib/kubelet/pods/4d54b307-a689-492f-99dd-9f7adb834498/volumes" Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.878263 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ea7c58b-3f82-4e10-9e02-12fde72999ca-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:42 crc kubenswrapper[4706]: I1127 07:26:42.878310 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shl5m\" (UniqueName: \"kubernetes.io/projected/1ea7c58b-3f82-4e10-9e02-12fde72999ca-kube-api-access-shl5m\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.251905 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-1" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.100:9292/healthcheck\": read tcp 10.217.0.2:47516->10.217.0.100:9292: read: connection reset by peer" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.251948 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-1" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.100:9292/healthcheck\": read tcp 10.217.0.2:47526->10.217.0.100:9292: read: connection reset by peer" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.261582 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" event={"ID":"1ea7c58b-3f82-4e10-9e02-12fde72999ca","Type":"ContainerDied","Data":"4eeef396afeb135bb15068f6cecc9085389e8f33668f51dbd47d65d4aa41e341"} Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.261615 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eeef396afeb135bb15068f6cecc9085389e8f33668f51dbd47d65d4aa41e341" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.261651 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb7e1-account-delete-7crsq" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.711047 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.716979 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891035 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-dev\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891432 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-run\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891461 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891485 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-run\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891506 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-dev\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891600 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-scripts\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891623 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-config-data\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891652 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-scripts\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891676 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-config-data\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891711 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-httpd-run\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891724 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-run" (OuterVolumeSpecName: "run") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891758 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-logs\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891785 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-var-locks-brick\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891814 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-logs\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891863 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-httpd-run\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891902 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-nvme\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891921 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891947 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-lib-modules\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891981 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-sys\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892009 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-lib-modules\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892056 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-iscsi\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892086 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-iscsi\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892118 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892154 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6xnk\" (UniqueName: \"kubernetes.io/projected/bd814f44-0460-400d-8570-594de21b5846-kube-api-access-j6xnk\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892198 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-nvme\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892262 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-var-locks-brick\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892322 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-sys\") pod \"bd814f44-0460-400d-8570-594de21b5846\" (UID: \"bd814f44-0460-400d-8570-594de21b5846\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892348 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cc4g\" (UniqueName: \"kubernetes.io/projected/18672d3a-3a06-4450-bad8-2048f1afc942-kube-api-access-5cc4g\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892373 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"18672d3a-3a06-4450-bad8-2048f1afc942\" (UID: \"18672d3a-3a06-4450-bad8-2048f1afc942\") " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892818 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891813 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-dev" (OuterVolumeSpecName: "dev") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894717 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.891845 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-run" (OuterVolumeSpecName: "run") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894730 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892855 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.892913 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-dev" (OuterVolumeSpecName: "dev") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894390 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894648 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-sys" (OuterVolumeSpecName: "sys") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894675 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894705 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894743 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.894728 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-logs" (OuterVolumeSpecName: "logs") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.895328 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-sys" (OuterVolumeSpecName: "sys") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.895671 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.895387 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.895990 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-logs" (OuterVolumeSpecName: "logs") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.896450 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.898526 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.898705 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.898878 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance-cache") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.898893 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-scripts" (OuterVolumeSpecName: "scripts") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.899463 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18672d3a-3a06-4450-bad8-2048f1afc942-kube-api-access-5cc4g" (OuterVolumeSpecName: "kube-api-access-5cc4g") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "kube-api-access-5cc4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.900281 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd814f44-0460-400d-8570-594de21b5846-kube-api-access-j6xnk" (OuterVolumeSpecName: "kube-api-access-j6xnk") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "kube-api-access-j6xnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.901011 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.902335 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-scripts" (OuterVolumeSpecName: "scripts") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.932988 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-config-data" (OuterVolumeSpecName: "config-data") pod "bd814f44-0460-400d-8570-594de21b5846" (UID: "bd814f44-0460-400d-8570-594de21b5846"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.933398 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-config-data" (OuterVolumeSpecName: "config-data") pod "18672d3a-3a06-4450-bad8-2048f1afc942" (UID: "18672d3a-3a06-4450-bad8-2048f1afc942"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998267 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998328 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998363 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18672d3a-3a06-4450-bad8-2048f1afc942-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998382 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998429 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998456 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998475 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998493 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998511 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998534 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998563 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998590 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6xnk\" (UniqueName: \"kubernetes.io/projected/bd814f44-0460-400d-8570-594de21b5846-kube-api-access-j6xnk\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998612 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998631 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998650 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cc4g\" (UniqueName: \"kubernetes.io/projected/18672d3a-3a06-4450-bad8-2048f1afc942-kube-api-access-5cc4g\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998678 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998715 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 07:26:43 crc kubenswrapper[4706]: I1127 07:26:43.998733 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/18672d3a-3a06-4450-bad8-2048f1afc942-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.002902 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.002959 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.002990 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd814f44-0460-400d-8570-594de21b5846-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.003008 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.003026 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18672d3a-3a06-4450-bad8-2048f1afc942-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.003044 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.003066 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd814f44-0460-400d-8570-594de21b5846-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.003084 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.003105 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd814f44-0460-400d-8570-594de21b5846-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.016700 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.022059 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.025800 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.026761 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.104311 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.104344 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.104357 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.104367 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.272038 4706 generic.go:334] "Generic (PLEG): container finished" podID="18672d3a-3a06-4450-bad8-2048f1afc942" containerID="21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05" exitCode=0 Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.272098 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"18672d3a-3a06-4450-bad8-2048f1afc942","Type":"ContainerDied","Data":"21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05"} Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.272124 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"18672d3a-3a06-4450-bad8-2048f1afc942","Type":"ContainerDied","Data":"6f50833712b942cf4a28966f15b3dea5d4d044ccc414cbe86c89403eafe3497c"} Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.272143 4706 scope.go:117] "RemoveContainer" containerID="21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.272266 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.278694 4706 generic.go:334] "Generic (PLEG): container finished" podID="bd814f44-0460-400d-8570-594de21b5846" containerID="fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280" exitCode=0 Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.278740 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"bd814f44-0460-400d-8570-594de21b5846","Type":"ContainerDied","Data":"fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280"} Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.278769 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"bd814f44-0460-400d-8570-594de21b5846","Type":"ContainerDied","Data":"06269cef2d4a387255f943064d9734d6c1d8b1cd7d4eb59f1de134f94a61c081"} Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.279168 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.301766 4706 scope.go:117] "RemoveContainer" containerID="5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.310549 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.320023 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.324939 4706 scope.go:117] "RemoveContainer" containerID="21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05" Nov 27 07:26:44 crc kubenswrapper[4706]: E1127 07:26:44.325453 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05\": container with ID starting with 21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05 not found: ID does not exist" containerID="21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.325477 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05"} err="failed to get container status \"21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05\": rpc error: code = NotFound desc = could not find container \"21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05\": container with ID starting with 21545a9eacaa1e263fad3c4c83240d6d5f2fe717b7c06462d6579e9fd7b15e05 not found: ID does not exist" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.325498 4706 scope.go:117] "RemoveContainer" containerID="5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584" Nov 27 07:26:44 crc kubenswrapper[4706]: E1127 07:26:44.325892 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584\": container with ID starting with 5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584 not found: ID does not exist" containerID="5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.325918 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584"} err="failed to get container status \"5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584\": rpc error: code = NotFound desc = could not find container \"5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584\": container with ID starting with 5fd4b76910ced9d560b3939212ed2e412f5becb4f4fe0929cc9276664e2c3584 not found: ID does not exist" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.325937 4706 scope.go:117] "RemoveContainer" containerID="fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.327364 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.337441 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.357418 4706 scope.go:117] "RemoveContainer" containerID="b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.374846 4706 scope.go:117] "RemoveContainer" containerID="fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280" Nov 27 07:26:44 crc kubenswrapper[4706]: E1127 07:26:44.375587 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280\": container with ID starting with fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280 not found: ID does not exist" containerID="fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.375671 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280"} err="failed to get container status \"fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280\": rpc error: code = NotFound desc = could not find container \"fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280\": container with ID starting with fb50eb48428390297c27d14a86398249b3d0adbd55341d4a297d97e455ad3280 not found: ID does not exist" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.375725 4706 scope.go:117] "RemoveContainer" containerID="b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651" Nov 27 07:26:44 crc kubenswrapper[4706]: E1127 07:26:44.376179 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651\": container with ID starting with b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651 not found: ID does not exist" containerID="b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.376211 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651"} err="failed to get container status \"b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651\": rpc error: code = NotFound desc = could not find container \"b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651\": container with ID starting with b6fb4d8fe6b6867a0d8c68ddd7dfabfeaae1fc9e4f71e7e36c1031361b023651 not found: ID does not exist" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.792342 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" path="/var/lib/kubelet/pods/18672d3a-3a06-4450-bad8-2048f1afc942/volumes" Nov 27 07:26:44 crc kubenswrapper[4706]: I1127 07:26:44.793818 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd814f44-0460-400d-8570-594de21b5846" path="/var/lib/kubelet/pods/bd814f44-0460-400d-8570-594de21b5846/volumes" Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.156522 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanceb7e1-account-delete-7crsq"] Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.163687 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-x9w4b"] Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.170783 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanceb7e1-account-delete-7crsq"] Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.177411 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.177469 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.177628 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd"] Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.183603 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-b7e1-account-create-update-2pkrd"] Nov 27 07:26:45 crc kubenswrapper[4706]: I1127 07:26:45.189754 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-x9w4b"] Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.701566 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-b961-account-create-update-n6p96"] Nov 27 07:26:46 crc kubenswrapper[4706]: E1127 07:26:46.702438 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d54b307-a689-492f-99dd-9f7adb834498" containerName="openstackclient" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.702523 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d54b307-a689-492f-99dd-9f7adb834498" containerName="openstackclient" Nov 27 07:26:46 crc kubenswrapper[4706]: E1127 07:26:46.702603 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-httpd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.702660 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-httpd" Nov 27 07:26:46 crc kubenswrapper[4706]: E1127 07:26:46.702793 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-httpd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.702851 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-httpd" Nov 27 07:26:46 crc kubenswrapper[4706]: E1127 07:26:46.702935 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-log" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.702993 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-log" Nov 27 07:26:46 crc kubenswrapper[4706]: E1127 07:26:46.703057 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-log" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703107 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-log" Nov 27 07:26:46 crc kubenswrapper[4706]: E1127 07:26:46.703158 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea7c58b-3f82-4e10-9e02-12fde72999ca" containerName="mariadb-account-delete" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703205 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea7c58b-3f82-4e10-9e02-12fde72999ca" containerName="mariadb-account-delete" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703414 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea7c58b-3f82-4e10-9e02-12fde72999ca" containerName="mariadb-account-delete" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703489 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d54b307-a689-492f-99dd-9f7adb834498" containerName="openstackclient" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703550 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-log" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703611 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-httpd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703666 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="18672d3a-3a06-4450-bad8-2048f1afc942" containerName="glance-log" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.703725 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd814f44-0460-400d-8570-594de21b5846" containerName="glance-httpd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.704240 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.707039 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-5pqfd"] Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.707097 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.707840 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.718122 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-b961-account-create-update-n6p96"] Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.743271 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc455bd-188c-4b31-83c4-db243d3cb16a-operator-scripts\") pod \"glance-b961-account-create-update-n6p96\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.743552 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/977c0fc4-5820-41e1-b45a-ba76b8068df6-operator-scripts\") pod \"glance-db-create-5pqfd\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.743685 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcj5z\" (UniqueName: \"kubernetes.io/projected/afc455bd-188c-4b31-83c4-db243d3cb16a-kube-api-access-hcj5z\") pod \"glance-b961-account-create-update-n6p96\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.743794 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tqm9\" (UniqueName: \"kubernetes.io/projected/977c0fc4-5820-41e1-b45a-ba76b8068df6-kube-api-access-4tqm9\") pod \"glance-db-create-5pqfd\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.747082 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-5pqfd"] Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.791419 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea7c58b-3f82-4e10-9e02-12fde72999ca" path="/var/lib/kubelet/pods/1ea7c58b-3f82-4e10-9e02-12fde72999ca/volumes" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.791991 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65045997-87c4-4ba0-b18a-9e656f61fc16" path="/var/lib/kubelet/pods/65045997-87c4-4ba0-b18a-9e656f61fc16/volumes" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.792583 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed44d2c-d0dc-4c85-b9ff-810139ace09b" path="/var/lib/kubelet/pods/aed44d2c-d0dc-4c85-b9ff-810139ace09b/volumes" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.845052 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcj5z\" (UniqueName: \"kubernetes.io/projected/afc455bd-188c-4b31-83c4-db243d3cb16a-kube-api-access-hcj5z\") pod \"glance-b961-account-create-update-n6p96\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.845106 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tqm9\" (UniqueName: \"kubernetes.io/projected/977c0fc4-5820-41e1-b45a-ba76b8068df6-kube-api-access-4tqm9\") pod \"glance-db-create-5pqfd\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.845150 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc455bd-188c-4b31-83c4-db243d3cb16a-operator-scripts\") pod \"glance-b961-account-create-update-n6p96\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.845316 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/977c0fc4-5820-41e1-b45a-ba76b8068df6-operator-scripts\") pod \"glance-db-create-5pqfd\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.846024 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/977c0fc4-5820-41e1-b45a-ba76b8068df6-operator-scripts\") pod \"glance-db-create-5pqfd\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.846087 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc455bd-188c-4b31-83c4-db243d3cb16a-operator-scripts\") pod \"glance-b961-account-create-update-n6p96\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.862022 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcj5z\" (UniqueName: \"kubernetes.io/projected/afc455bd-188c-4b31-83c4-db243d3cb16a-kube-api-access-hcj5z\") pod \"glance-b961-account-create-update-n6p96\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:46 crc kubenswrapper[4706]: I1127 07:26:46.862439 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tqm9\" (UniqueName: \"kubernetes.io/projected/977c0fc4-5820-41e1-b45a-ba76b8068df6-kube-api-access-4tqm9\") pod \"glance-db-create-5pqfd\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:47 crc kubenswrapper[4706]: I1127 07:26:47.031333 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:47 crc kubenswrapper[4706]: I1127 07:26:47.043014 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:47 crc kubenswrapper[4706]: I1127 07:26:47.509329 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-b961-account-create-update-n6p96"] Nov 27 07:26:47 crc kubenswrapper[4706]: I1127 07:26:47.536476 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-5pqfd"] Nov 27 07:26:47 crc kubenswrapper[4706]: W1127 07:26:47.538701 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod977c0fc4_5820_41e1_b45a_ba76b8068df6.slice/crio-f7e991f5d94a49621b6acbaa7344c8974396ae7416d3a97ef64d55ffc4457936 WatchSource:0}: Error finding container f7e991f5d94a49621b6acbaa7344c8974396ae7416d3a97ef64d55ffc4457936: Status 404 returned error can't find the container with id f7e991f5d94a49621b6acbaa7344c8974396ae7416d3a97ef64d55ffc4457936 Nov 27 07:26:48 crc kubenswrapper[4706]: I1127 07:26:48.313687 4706 generic.go:334] "Generic (PLEG): container finished" podID="afc455bd-188c-4b31-83c4-db243d3cb16a" containerID="3e278089af7f8ca36bd2f8e0b74d74720499b1eac07ce3f65f93eb82f4a59b4c" exitCode=0 Nov 27 07:26:48 crc kubenswrapper[4706]: I1127 07:26:48.313746 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" event={"ID":"afc455bd-188c-4b31-83c4-db243d3cb16a","Type":"ContainerDied","Data":"3e278089af7f8ca36bd2f8e0b74d74720499b1eac07ce3f65f93eb82f4a59b4c"} Nov 27 07:26:48 crc kubenswrapper[4706]: I1127 07:26:48.314102 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" event={"ID":"afc455bd-188c-4b31-83c4-db243d3cb16a","Type":"ContainerStarted","Data":"f4e07695a877c2131f45c4eed140261a127dc4de07c982a119fd0e1cf8a321dd"} Nov 27 07:26:48 crc kubenswrapper[4706]: I1127 07:26:48.315807 4706 generic.go:334] "Generic (PLEG): container finished" podID="977c0fc4-5820-41e1-b45a-ba76b8068df6" containerID="ab5487f54e864099c15b01429a9a5faeffb0301a91c0bf59f0b9441b7143d40a" exitCode=0 Nov 27 07:26:48 crc kubenswrapper[4706]: I1127 07:26:48.315834 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-5pqfd" event={"ID":"977c0fc4-5820-41e1-b45a-ba76b8068df6","Type":"ContainerDied","Data":"ab5487f54e864099c15b01429a9a5faeffb0301a91c0bf59f0b9441b7143d40a"} Nov 27 07:26:48 crc kubenswrapper[4706]: I1127 07:26:48.315854 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-5pqfd" event={"ID":"977c0fc4-5820-41e1-b45a-ba76b8068df6","Type":"ContainerStarted","Data":"f7e991f5d94a49621b6acbaa7344c8974396ae7416d3a97ef64d55ffc4457936"} Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.788556 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.793903 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.798875 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcj5z\" (UniqueName: \"kubernetes.io/projected/afc455bd-188c-4b31-83c4-db243d3cb16a-kube-api-access-hcj5z\") pod \"afc455bd-188c-4b31-83c4-db243d3cb16a\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.807540 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc455bd-188c-4b31-83c4-db243d3cb16a-kube-api-access-hcj5z" (OuterVolumeSpecName: "kube-api-access-hcj5z") pod "afc455bd-188c-4b31-83c4-db243d3cb16a" (UID: "afc455bd-188c-4b31-83c4-db243d3cb16a"). InnerVolumeSpecName "kube-api-access-hcj5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.903909 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/977c0fc4-5820-41e1-b45a-ba76b8068df6-operator-scripts\") pod \"977c0fc4-5820-41e1-b45a-ba76b8068df6\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.903974 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tqm9\" (UniqueName: \"kubernetes.io/projected/977c0fc4-5820-41e1-b45a-ba76b8068df6-kube-api-access-4tqm9\") pod \"977c0fc4-5820-41e1-b45a-ba76b8068df6\" (UID: \"977c0fc4-5820-41e1-b45a-ba76b8068df6\") " Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.904016 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc455bd-188c-4b31-83c4-db243d3cb16a-operator-scripts\") pod \"afc455bd-188c-4b31-83c4-db243d3cb16a\" (UID: \"afc455bd-188c-4b31-83c4-db243d3cb16a\") " Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.904404 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcj5z\" (UniqueName: \"kubernetes.io/projected/afc455bd-188c-4b31-83c4-db243d3cb16a-kube-api-access-hcj5z\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.904498 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977c0fc4-5820-41e1-b45a-ba76b8068df6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "977c0fc4-5820-41e1-b45a-ba76b8068df6" (UID: "977c0fc4-5820-41e1-b45a-ba76b8068df6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.904632 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afc455bd-188c-4b31-83c4-db243d3cb16a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afc455bd-188c-4b31-83c4-db243d3cb16a" (UID: "afc455bd-188c-4b31-83c4-db243d3cb16a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:26:49 crc kubenswrapper[4706]: I1127 07:26:49.910466 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/977c0fc4-5820-41e1-b45a-ba76b8068df6-kube-api-access-4tqm9" (OuterVolumeSpecName: "kube-api-access-4tqm9") pod "977c0fc4-5820-41e1-b45a-ba76b8068df6" (UID: "977c0fc4-5820-41e1-b45a-ba76b8068df6"). InnerVolumeSpecName "kube-api-access-4tqm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.005773 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/977c0fc4-5820-41e1-b45a-ba76b8068df6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.005808 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tqm9\" (UniqueName: \"kubernetes.io/projected/977c0fc4-5820-41e1-b45a-ba76b8068df6-kube-api-access-4tqm9\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.005818 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc455bd-188c-4b31-83c4-db243d3cb16a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.339592 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.339581 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b961-account-create-update-n6p96" event={"ID":"afc455bd-188c-4b31-83c4-db243d3cb16a","Type":"ContainerDied","Data":"f4e07695a877c2131f45c4eed140261a127dc4de07c982a119fd0e1cf8a321dd"} Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.339891 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4e07695a877c2131f45c4eed140261a127dc4de07c982a119fd0e1cf8a321dd" Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.342402 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-5pqfd" event={"ID":"977c0fc4-5820-41e1-b45a-ba76b8068df6","Type":"ContainerDied","Data":"f7e991f5d94a49621b6acbaa7344c8974396ae7416d3a97ef64d55ffc4457936"} Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.342453 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7e991f5d94a49621b6acbaa7344c8974396ae7416d3a97ef64d55ffc4457936" Nov 27 07:26:50 crc kubenswrapper[4706]: I1127 07:26:50.342559 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-5pqfd" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.830440 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-8mzkw"] Nov 27 07:26:51 crc kubenswrapper[4706]: E1127 07:26:51.831526 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc455bd-188c-4b31-83c4-db243d3cb16a" containerName="mariadb-account-create-update" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.831547 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc455bd-188c-4b31-83c4-db243d3cb16a" containerName="mariadb-account-create-update" Nov 27 07:26:51 crc kubenswrapper[4706]: E1127 07:26:51.831578 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977c0fc4-5820-41e1-b45a-ba76b8068df6" containerName="mariadb-database-create" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.831588 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="977c0fc4-5820-41e1-b45a-ba76b8068df6" containerName="mariadb-database-create" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.831794 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc455bd-188c-4b31-83c4-db243d3cb16a" containerName="mariadb-account-create-update" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.831830 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="977c0fc4-5820-41e1-b45a-ba76b8068df6" containerName="mariadb-database-create" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.832530 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.835350 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.835533 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.835811 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-skmkp" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.840556 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-8mzkw"] Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.937577 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-config-data\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.937637 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-combined-ca-bundle\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.937699 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-db-sync-config-data\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:51 crc kubenswrapper[4706]: I1127 07:26:51.937762 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9n8t\" (UniqueName: \"kubernetes.io/projected/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-kube-api-access-h9n8t\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.039381 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-config-data\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.039436 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-combined-ca-bundle\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.039462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-db-sync-config-data\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.039508 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9n8t\" (UniqueName: \"kubernetes.io/projected/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-kube-api-access-h9n8t\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.044715 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-combined-ca-bundle\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.044800 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-config-data\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.044976 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-db-sync-config-data\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.080992 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9n8t\" (UniqueName: \"kubernetes.io/projected/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-kube-api-access-h9n8t\") pod \"glance-db-sync-8mzkw\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.147661 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:52 crc kubenswrapper[4706]: I1127 07:26:52.553393 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-8mzkw"] Nov 27 07:26:53 crc kubenswrapper[4706]: I1127 07:26:53.367930 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8mzkw" event={"ID":"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb","Type":"ContainerStarted","Data":"12fa7d40f5a433302eddef5bd9fb154643de88e96eb951204d94bd7482fac427"} Nov 27 07:26:53 crc kubenswrapper[4706]: I1127 07:26:53.368004 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8mzkw" event={"ID":"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb","Type":"ContainerStarted","Data":"a0f29fbfea5186f5f212903df0746102bdbbeedce5ba882e13d9d2f0239c5155"} Nov 27 07:26:53 crc kubenswrapper[4706]: I1127 07:26:53.390177 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-8mzkw" podStartSLOduration=2.390146832 podStartE2EDuration="2.390146832s" podCreationTimestamp="2025-11-27 07:26:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:26:53.382185763 +0000 UTC m=+1097.271776603" watchObservedRunningTime="2025-11-27 07:26:53.390146832 +0000 UTC m=+1097.279737682" Nov 27 07:26:56 crc kubenswrapper[4706]: I1127 07:26:56.394088 4706 generic.go:334] "Generic (PLEG): container finished" podID="895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" containerID="12fa7d40f5a433302eddef5bd9fb154643de88e96eb951204d94bd7482fac427" exitCode=0 Nov 27 07:26:56 crc kubenswrapper[4706]: I1127 07:26:56.394265 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8mzkw" event={"ID":"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb","Type":"ContainerDied","Data":"12fa7d40f5a433302eddef5bd9fb154643de88e96eb951204d94bd7482fac427"} Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.726132 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.922250 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-config-data\") pod \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.922387 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9n8t\" (UniqueName: \"kubernetes.io/projected/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-kube-api-access-h9n8t\") pod \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.922418 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-db-sync-config-data\") pod \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.922454 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-combined-ca-bundle\") pod \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\" (UID: \"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb\") " Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.928774 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" (UID: "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.929427 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-kube-api-access-h9n8t" (OuterVolumeSpecName: "kube-api-access-h9n8t") pod "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" (UID: "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb"). InnerVolumeSpecName "kube-api-access-h9n8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.957905 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" (UID: "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:57 crc kubenswrapper[4706]: I1127 07:26:57.963616 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-config-data" (OuterVolumeSpecName: "config-data") pod "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" (UID: "895ededf-d1dd-40d0-a9bb-ffca3ffe07fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.024784 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.024817 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9n8t\" (UniqueName: \"kubernetes.io/projected/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-kube-api-access-h9n8t\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.024827 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.024835 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.411477 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8mzkw" event={"ID":"895ededf-d1dd-40d0-a9bb-ffca3ffe07fb","Type":"ContainerDied","Data":"a0f29fbfea5186f5f212903df0746102bdbbeedce5ba882e13d9d2f0239c5155"} Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.411518 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0f29fbfea5186f5f212903df0746102bdbbeedce5ba882e13d9d2f0239c5155" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.411586 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8mzkw" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.738246 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:26:58 crc kubenswrapper[4706]: E1127 07:26:58.738858 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" containerName="glance-db-sync" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.738873 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" containerName="glance-db-sync" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.739030 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" containerName="glance-db-sync" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.739861 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.742001 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.742296 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.742915 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.743247 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.743420 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-skmkp" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.743583 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.752005 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.933923 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.933977 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-config-data\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.934099 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.934278 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-logs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.934316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.934394 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh5w8\" (UniqueName: \"kubernetes.io/projected/5d61ddd1-5738-4a88-933c-1f59c651138a-kube-api-access-bh5w8\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.934464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-httpd-run\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.934561 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-scripts\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:58 crc kubenswrapper[4706]: I1127 07:26:58.934614 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036298 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-logs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036352 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036386 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh5w8\" (UniqueName: \"kubernetes.io/projected/5d61ddd1-5738-4a88-933c-1f59c651138a-kube-api-access-bh5w8\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036410 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-httpd-run\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036449 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-scripts\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036469 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036510 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036532 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-config-data\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036566 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.036883 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.037098 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-logs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.038954 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-httpd-run\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.041733 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.052135 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.052166 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.052760 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-scripts\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.053174 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-config-data\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.055918 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.060668 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh5w8\" (UniqueName: \"kubernetes.io/projected/5d61ddd1-5738-4a88-933c-1f59c651138a-kube-api-access-bh5w8\") pod \"glance-default-single-0\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:26:59 crc kubenswrapper[4706]: I1127 07:26:59.360308 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:00 crc kubenswrapper[4706]: I1127 07:26:59.646779 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:00 crc kubenswrapper[4706]: I1127 07:26:59.882112 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:00 crc kubenswrapper[4706]: I1127 07:27:00.430024 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"5d61ddd1-5738-4a88-933c-1f59c651138a","Type":"ContainerStarted","Data":"5a747c77602de93ef6a5d9cd8bf4e8f613fb668212608b965af9c060f6651cf8"} Nov 27 07:27:01 crc kubenswrapper[4706]: I1127 07:27:01.444582 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"5d61ddd1-5738-4a88-933c-1f59c651138a","Type":"ContainerStarted","Data":"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d"} Nov 27 07:27:01 crc kubenswrapper[4706]: I1127 07:27:01.445134 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"5d61ddd1-5738-4a88-933c-1f59c651138a","Type":"ContainerStarted","Data":"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059"} Nov 27 07:27:01 crc kubenswrapper[4706]: I1127 07:27:01.444764 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-log" containerID="cri-o://40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059" gracePeriod=30 Nov 27 07:27:01 crc kubenswrapper[4706]: I1127 07:27:01.444814 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-httpd" containerID="cri-o://1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d" gracePeriod=30 Nov 27 07:27:01 crc kubenswrapper[4706]: I1127 07:27:01.471054 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.471038174 podStartE2EDuration="3.471038174s" podCreationTimestamp="2025-11-27 07:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:01.465828195 +0000 UTC m=+1105.355419005" watchObservedRunningTime="2025-11-27 07:27:01.471038174 +0000 UTC m=+1105.360628984" Nov 27 07:27:01 crc kubenswrapper[4706]: I1127 07:27:01.927380 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081015 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-logs\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081056 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-internal-tls-certs\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081098 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-public-tls-certs\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081116 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-httpd-run\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081150 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-combined-ca-bundle\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081168 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081189 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-scripts\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081235 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-config-data\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081288 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh5w8\" (UniqueName: \"kubernetes.io/projected/5d61ddd1-5738-4a88-933c-1f59c651138a-kube-api-access-bh5w8\") pod \"5d61ddd1-5738-4a88-933c-1f59c651138a\" (UID: \"5d61ddd1-5738-4a88-933c-1f59c651138a\") " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081668 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-logs" (OuterVolumeSpecName: "logs") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.081715 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.085634 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d61ddd1-5738-4a88-933c-1f59c651138a-kube-api-access-bh5w8" (OuterVolumeSpecName: "kube-api-access-bh5w8") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "kube-api-access-bh5w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.086152 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.094360 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-scripts" (OuterVolumeSpecName: "scripts") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.116200 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.117884 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.121614 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.134818 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-config-data" (OuterVolumeSpecName: "config-data") pod "5d61ddd1-5738-4a88-933c-1f59c651138a" (UID: "5d61ddd1-5738-4a88-933c-1f59c651138a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183163 4706 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183199 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183210 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183268 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183287 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183299 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183310 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh5w8\" (UniqueName: \"kubernetes.io/projected/5d61ddd1-5738-4a88-933c-1f59c651138a-kube-api-access-bh5w8\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183325 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d61ddd1-5738-4a88-933c-1f59c651138a-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.183337 4706 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d61ddd1-5738-4a88-933c-1f59c651138a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.207890 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.285096 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.456208 4706 generic.go:334] "Generic (PLEG): container finished" podID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerID="1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d" exitCode=0 Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.456268 4706 generic.go:334] "Generic (PLEG): container finished" podID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerID="40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059" exitCode=143 Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.456294 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"5d61ddd1-5738-4a88-933c-1f59c651138a","Type":"ContainerDied","Data":"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d"} Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.456352 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"5d61ddd1-5738-4a88-933c-1f59c651138a","Type":"ContainerDied","Data":"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059"} Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.456367 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"5d61ddd1-5738-4a88-933c-1f59c651138a","Type":"ContainerDied","Data":"5a747c77602de93ef6a5d9cd8bf4e8f613fb668212608b965af9c060f6651cf8"} Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.456364 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.456386 4706 scope.go:117] "RemoveContainer" containerID="1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.498179 4706 scope.go:117] "RemoveContainer" containerID="40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.499844 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.510632 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.532149 4706 scope.go:117] "RemoveContainer" containerID="1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d" Nov 27 07:27:02 crc kubenswrapper[4706]: E1127 07:27:02.533043 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d\": container with ID starting with 1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d not found: ID does not exist" containerID="1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.533142 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d"} err="failed to get container status \"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d\": rpc error: code = NotFound desc = could not find container \"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d\": container with ID starting with 1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d not found: ID does not exist" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.533171 4706 scope.go:117] "RemoveContainer" containerID="40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059" Nov 27 07:27:02 crc kubenswrapper[4706]: E1127 07:27:02.533656 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059\": container with ID starting with 40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059 not found: ID does not exist" containerID="40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.533688 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059"} err="failed to get container status \"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059\": rpc error: code = NotFound desc = could not find container \"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059\": container with ID starting with 40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059 not found: ID does not exist" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.533708 4706 scope.go:117] "RemoveContainer" containerID="1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.534743 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:02 crc kubenswrapper[4706]: E1127 07:27:02.535186 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-log" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.535215 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-log" Nov 27 07:27:02 crc kubenswrapper[4706]: E1127 07:27:02.535270 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-httpd" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.535284 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-httpd" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.535573 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-log" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.535622 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" containerName="glance-httpd" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.536121 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d"} err="failed to get container status \"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d\": rpc error: code = NotFound desc = could not find container \"1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d\": container with ID starting with 1814dfc39d132b294a25cd3b5db0ede87d89e38a7b370db53702e3c1301d483d not found: ID does not exist" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.536187 4706 scope.go:117] "RemoveContainer" containerID="40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.537169 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059"} err="failed to get container status \"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059\": rpc error: code = NotFound desc = could not find container \"40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059\": container with ID starting with 40a0736a9a8e84f0bf759a9d75ee1189bd72946eada619b57ee16bb60534d059 not found: ID does not exist" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.537175 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.541458 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.541673 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.541729 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.541682 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.541921 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.544017 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.544122 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-skmkp" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.690691 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691003 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691133 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691255 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691363 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-logs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691473 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691581 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691664 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-httpd-run\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.691763 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm8qw\" (UniqueName: \"kubernetes.io/projected/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-kube-api-access-gm8qw\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.785241 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d61ddd1-5738-4a88-933c-1f59c651138a" path="/var/lib/kubelet/pods/5d61ddd1-5738-4a88-933c-1f59c651138a/volumes" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793495 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793533 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-logs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793562 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793583 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793599 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-httpd-run\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793620 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm8qw\" (UniqueName: \"kubernetes.io/projected/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-kube-api-access-gm8qw\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793640 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793697 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.793715 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.794546 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-httpd-run\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.794732 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-logs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.794741 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.797913 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.799395 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.800153 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.804853 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.806578 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.813607 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.821567 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm8qw\" (UniqueName: \"kubernetes.io/projected/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-kube-api-access-gm8qw\") pod \"glance-default-single-0\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:02 crc kubenswrapper[4706]: I1127 07:27:02.899089 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:03 crc kubenswrapper[4706]: I1127 07:27:03.368445 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:03 crc kubenswrapper[4706]: W1127 07:27:03.379229 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb09b59ae_48b2_4fda_80f3_591c6ce27dcd.slice/crio-259b55df6b30aa142d7513fbc67900f5075000aadbfed859fb09f926379979df WatchSource:0}: Error finding container 259b55df6b30aa142d7513fbc67900f5075000aadbfed859fb09f926379979df: Status 404 returned error can't find the container with id 259b55df6b30aa142d7513fbc67900f5075000aadbfed859fb09f926379979df Nov 27 07:27:03 crc kubenswrapper[4706]: I1127 07:27:03.467459 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b09b59ae-48b2-4fda-80f3-591c6ce27dcd","Type":"ContainerStarted","Data":"259b55df6b30aa142d7513fbc67900f5075000aadbfed859fb09f926379979df"} Nov 27 07:27:04 crc kubenswrapper[4706]: I1127 07:27:04.476554 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b09b59ae-48b2-4fda-80f3-591c6ce27dcd","Type":"ContainerStarted","Data":"03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59"} Nov 27 07:27:04 crc kubenswrapper[4706]: I1127 07:27:04.476879 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b09b59ae-48b2-4fda-80f3-591c6ce27dcd","Type":"ContainerStarted","Data":"da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df"} Nov 27 07:27:04 crc kubenswrapper[4706]: I1127 07:27:04.505325 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.505310611 podStartE2EDuration="2.505310611s" podCreationTimestamp="2025-11-27 07:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:04.499747019 +0000 UTC m=+1108.389337829" watchObservedRunningTime="2025-11-27 07:27:04.505310611 +0000 UTC m=+1108.394901421" Nov 27 07:27:12 crc kubenswrapper[4706]: I1127 07:27:12.900424 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:12 crc kubenswrapper[4706]: I1127 07:27:12.902811 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:12 crc kubenswrapper[4706]: I1127 07:27:12.950370 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:12 crc kubenswrapper[4706]: I1127 07:27:12.959355 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:13 crc kubenswrapper[4706]: I1127 07:27:13.562596 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:13 crc kubenswrapper[4706]: I1127 07:27:13.562633 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.177687 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.178072 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.178145 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.179364 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d7703c2342c25e6f9c346c6d641fa2f157694340cd2c97e78820343e2f8c3744"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.179472 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://d7703c2342c25e6f9c346c6d641fa2f157694340cd2c97e78820343e2f8c3744" gracePeriod=600 Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.529374 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.537968 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.581141 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="d7703c2342c25e6f9c346c6d641fa2f157694340cd2c97e78820343e2f8c3744" exitCode=0 Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.581874 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"d7703c2342c25e6f9c346c6d641fa2f157694340cd2c97e78820343e2f8c3744"} Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.581906 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"6fa6ce74ad58cab1c7d76467a227f52ada0afc016b50f3d7c2b5893e9773b0ee"} Nov 27 07:27:15 crc kubenswrapper[4706]: I1127 07:27:15.581923 4706 scope.go:117] "RemoveContainer" containerID="db4ade098e79285766f3d1f1fb101ceeccbe33f1280ddb92a2252c3b8d5a2939" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.670366 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-8mzkw"] Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.675745 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-8mzkw"] Nov 27 07:27:16 crc kubenswrapper[4706]: E1127 07:27:16.721318 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 27 07:27:16 crc kubenswrapper[4706]: E1127 07:27:16.721392 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:17.221374967 +0000 UTC m=+1121.110965777 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-default-single-config-data" not found Nov 27 07:27:16 crc kubenswrapper[4706]: E1127 07:27:16.721614 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 27 07:27:16 crc kubenswrapper[4706]: E1127 07:27:16.721644 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:17.221636355 +0000 UTC m=+1121.111227165 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-scripts" not found Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.745275 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.754643 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanceb961-account-delete-8j9b7"] Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.755500 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.762964 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceb961-account-delete-8j9b7"] Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.784739 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="895ededf-d1dd-40d0-a9bb-ffca3ffe07fb" path="/var/lib/kubelet/pods/895ededf-d1dd-40d0-a9bb-ffca3ffe07fb/volumes" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.822750 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3895b57f-2dad-4d55-8818-e0c96b975933-operator-scripts\") pod \"glanceb961-account-delete-8j9b7\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.822794 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crptz\" (UniqueName: \"kubernetes.io/projected/3895b57f-2dad-4d55-8818-e0c96b975933-kube-api-access-crptz\") pod \"glanceb961-account-delete-8j9b7\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.924653 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3895b57f-2dad-4d55-8818-e0c96b975933-operator-scripts\") pod \"glanceb961-account-delete-8j9b7\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.924716 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crptz\" (UniqueName: \"kubernetes.io/projected/3895b57f-2dad-4d55-8818-e0c96b975933-kube-api-access-crptz\") pod \"glanceb961-account-delete-8j9b7\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.926489 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3895b57f-2dad-4d55-8818-e0c96b975933-operator-scripts\") pod \"glanceb961-account-delete-8j9b7\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:16 crc kubenswrapper[4706]: I1127 07:27:16.952044 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crptz\" (UniqueName: \"kubernetes.io/projected/3895b57f-2dad-4d55-8818-e0c96b975933-kube-api-access-crptz\") pod \"glanceb961-account-delete-8j9b7\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:17 crc kubenswrapper[4706]: I1127 07:27:17.079839 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:17 crc kubenswrapper[4706]: E1127 07:27:17.228727 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 27 07:27:17 crc kubenswrapper[4706]: E1127 07:27:17.229026 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:18.229011477 +0000 UTC m=+1122.118602287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-scripts" not found Nov 27 07:27:17 crc kubenswrapper[4706]: E1127 07:27:17.229443 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 27 07:27:17 crc kubenswrapper[4706]: E1127 07:27:17.229480 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:18.229470422 +0000 UTC m=+1122.119061232 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-default-single-config-data" not found Nov 27 07:27:17 crc kubenswrapper[4706]: I1127 07:27:17.491927 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceb961-account-delete-8j9b7"] Nov 27 07:27:17 crc kubenswrapper[4706]: W1127 07:27:17.492978 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3895b57f_2dad_4d55_8818_e0c96b975933.slice/crio-e11763f9c0262e56451b803d19d1b8b13e377823970e6cfeffaae090b3176f40 WatchSource:0}: Error finding container e11763f9c0262e56451b803d19d1b8b13e377823970e6cfeffaae090b3176f40: Status 404 returned error can't find the container with id e11763f9c0262e56451b803d19d1b8b13e377823970e6cfeffaae090b3176f40 Nov 27 07:27:17 crc kubenswrapper[4706]: I1127 07:27:17.599872 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" event={"ID":"3895b57f-2dad-4d55-8818-e0c96b975933","Type":"ContainerStarted","Data":"e11763f9c0262e56451b803d19d1b8b13e377823970e6cfeffaae090b3176f40"} Nov 27 07:27:17 crc kubenswrapper[4706]: I1127 07:27:17.600034 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-log" containerID="cri-o://da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df" gracePeriod=30 Nov 27 07:27:17 crc kubenswrapper[4706]: I1127 07:27:17.600148 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-httpd" containerID="cri-o://03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59" gracePeriod=30 Nov 27 07:27:18 crc kubenswrapper[4706]: E1127 07:27:18.241817 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 27 07:27:18 crc kubenswrapper[4706]: E1127 07:27:18.241881 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:20.241867478 +0000 UTC m=+1124.131458288 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-scripts" not found Nov 27 07:27:18 crc kubenswrapper[4706]: E1127 07:27:18.241938 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 27 07:27:18 crc kubenswrapper[4706]: E1127 07:27:18.242034 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:20.242012712 +0000 UTC m=+1124.131603602 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-default-single-config-data" not found Nov 27 07:27:18 crc kubenswrapper[4706]: I1127 07:27:18.610575 4706 generic.go:334] "Generic (PLEG): container finished" podID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerID="da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df" exitCode=143 Nov 27 07:27:18 crc kubenswrapper[4706]: I1127 07:27:18.610675 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b09b59ae-48b2-4fda-80f3-591c6ce27dcd","Type":"ContainerDied","Data":"da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df"} Nov 27 07:27:18 crc kubenswrapper[4706]: I1127 07:27:18.613207 4706 generic.go:334] "Generic (PLEG): container finished" podID="3895b57f-2dad-4d55-8818-e0c96b975933" containerID="d13c12fb2ce58f6e2af465403ff0a52ce1dd6f62137f267bbcec4156b3b602b3" exitCode=0 Nov 27 07:27:18 crc kubenswrapper[4706]: I1127 07:27:18.613255 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" event={"ID":"3895b57f-2dad-4d55-8818-e0c96b975933","Type":"ContainerDied","Data":"d13c12fb2ce58f6e2af465403ff0a52ce1dd6f62137f267bbcec4156b3b602b3"} Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.023701 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.072254 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3895b57f-2dad-4d55-8818-e0c96b975933-operator-scripts\") pod \"3895b57f-2dad-4d55-8818-e0c96b975933\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.072302 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crptz\" (UniqueName: \"kubernetes.io/projected/3895b57f-2dad-4d55-8818-e0c96b975933-kube-api-access-crptz\") pod \"3895b57f-2dad-4d55-8818-e0c96b975933\" (UID: \"3895b57f-2dad-4d55-8818-e0c96b975933\") " Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.073040 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3895b57f-2dad-4d55-8818-e0c96b975933-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3895b57f-2dad-4d55-8818-e0c96b975933" (UID: "3895b57f-2dad-4d55-8818-e0c96b975933"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.073659 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3895b57f-2dad-4d55-8818-e0c96b975933-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.081290 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3895b57f-2dad-4d55-8818-e0c96b975933-kube-api-access-crptz" (OuterVolumeSpecName: "kube-api-access-crptz") pod "3895b57f-2dad-4d55-8818-e0c96b975933" (UID: "3895b57f-2dad-4d55-8818-e0c96b975933"). InnerVolumeSpecName "kube-api-access-crptz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.174480 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crptz\" (UniqueName: \"kubernetes.io/projected/3895b57f-2dad-4d55-8818-e0c96b975933-kube-api-access-crptz\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:20 crc kubenswrapper[4706]: E1127 07:27:20.275246 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 27 07:27:20 crc kubenswrapper[4706]: E1127 07:27:20.275313 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:24.275295352 +0000 UTC m=+1128.164886162 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-default-single-config-data" not found Nov 27 07:27:20 crc kubenswrapper[4706]: E1127 07:27:20.275383 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 27 07:27:20 crc kubenswrapper[4706]: E1127 07:27:20.275493 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts podName:b09b59ae-48b2-4fda-80f3-591c6ce27dcd nodeName:}" failed. No retries permitted until 2025-11-27 07:27:24.275458787 +0000 UTC m=+1128.165049647 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts") pod "glance-default-single-0" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd") : secret "glance-scripts" not found Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.630053 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" event={"ID":"3895b57f-2dad-4d55-8818-e0c96b975933","Type":"ContainerDied","Data":"e11763f9c0262e56451b803d19d1b8b13e377823970e6cfeffaae090b3176f40"} Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.630105 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e11763f9c0262e56451b803d19d1b8b13e377823970e6cfeffaae090b3176f40" Nov 27 07:27:20 crc kubenswrapper[4706]: I1127 07:27:20.630162 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb961-account-delete-8j9b7" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.270323 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.393163 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-logs\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.393258 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm8qw\" (UniqueName: \"kubernetes.io/projected/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-kube-api-access-gm8qw\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.393310 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-public-tls-certs\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.393398 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-combined-ca-bundle\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.393439 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.393834 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-logs" (OuterVolumeSpecName: "logs") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.394113 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-internal-tls-certs\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.394178 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.394250 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.394293 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-httpd-run\") pod \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\" (UID: \"b09b59ae-48b2-4fda-80f3-591c6ce27dcd\") " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.394625 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.395046 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.398445 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts" (OuterVolumeSpecName: "scripts") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.398740 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.398909 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-kube-api-access-gm8qw" (OuterVolumeSpecName: "kube-api-access-gm8qw") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "kube-api-access-gm8qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.414341 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.428970 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.436338 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data" (OuterVolumeSpecName: "config-data") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.444640 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b09b59ae-48b2-4fda-80f3-591c6ce27dcd" (UID: "b09b59ae-48b2-4fda-80f3-591c6ce27dcd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496427 4706 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496470 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496488 4706 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496526 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496540 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496551 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496563 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm8qw\" (UniqueName: \"kubernetes.io/projected/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-kube-api-access-gm8qw\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.496577 4706 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b09b59ae-48b2-4fda-80f3-591c6ce27dcd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.513282 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.598357 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.640115 4706 generic.go:334] "Generic (PLEG): container finished" podID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerID="03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59" exitCode=0 Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.640339 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.640363 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b09b59ae-48b2-4fda-80f3-591c6ce27dcd","Type":"ContainerDied","Data":"03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59"} Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.641082 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b09b59ae-48b2-4fda-80f3-591c6ce27dcd","Type":"ContainerDied","Data":"259b55df6b30aa142d7513fbc67900f5075000aadbfed859fb09f926379979df"} Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.641108 4706 scope.go:117] "RemoveContainer" containerID="03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.667635 4706 scope.go:117] "RemoveContainer" containerID="da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.685123 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.701256 4706 scope.go:117] "RemoveContainer" containerID="03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59" Nov 27 07:27:21 crc kubenswrapper[4706]: E1127 07:27:21.701809 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59\": container with ID starting with 03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59 not found: ID does not exist" containerID="03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.701861 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59"} err="failed to get container status \"03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59\": rpc error: code = NotFound desc = could not find container \"03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59\": container with ID starting with 03b325e0a59a9ff8f601d609d57b02489e4d1123a36b1dc6b88ffebd2f7d7c59 not found: ID does not exist" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.701890 4706 scope.go:117] "RemoveContainer" containerID="da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df" Nov 27 07:27:21 crc kubenswrapper[4706]: E1127 07:27:21.702366 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df\": container with ID starting with da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df not found: ID does not exist" containerID="da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.702390 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df"} err="failed to get container status \"da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df\": rpc error: code = NotFound desc = could not find container \"da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df\": container with ID starting with da17b5dadbf8a8b16002d1d018129c15d6554df17afd5d4943d15138d28136df not found: ID does not exist" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.711155 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.785067 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-5pqfd"] Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.792638 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-5pqfd"] Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.799139 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-b961-account-create-update-n6p96"] Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.806150 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanceb961-account-delete-8j9b7"] Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.814683 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanceb961-account-delete-8j9b7"] Nov 27 07:27:21 crc kubenswrapper[4706]: E1127 07:27:21.820787 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb09b59ae_48b2_4fda_80f3_591c6ce27dcd.slice\": RecentStats: unable to find data in memory cache]" Nov 27 07:27:21 crc kubenswrapper[4706]: I1127 07:27:21.823569 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-b961-account-create-update-n6p96"] Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.813246 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3895b57f-2dad-4d55-8818-e0c96b975933" path="/var/lib/kubelet/pods/3895b57f-2dad-4d55-8818-e0c96b975933/volumes" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.813861 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="977c0fc4-5820-41e1-b45a-ba76b8068df6" path="/var/lib/kubelet/pods/977c0fc4-5820-41e1-b45a-ba76b8068df6/volumes" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.814528 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc455bd-188c-4b31-83c4-db243d3cb16a" path="/var/lib/kubelet/pods/afc455bd-188c-4b31-83c4-db243d3cb16a/volumes" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.815855 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" path="/var/lib/kubelet/pods/b09b59ae-48b2-4fda-80f3-591c6ce27dcd/volumes" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.819885 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-q9d6s"] Nov 27 07:27:22 crc kubenswrapper[4706]: E1127 07:27:22.820149 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-log" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.820167 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-log" Nov 27 07:27:22 crc kubenswrapper[4706]: E1127 07:27:22.820186 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-httpd" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.820192 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-httpd" Nov 27 07:27:22 crc kubenswrapper[4706]: E1127 07:27:22.820211 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3895b57f-2dad-4d55-8818-e0c96b975933" containerName="mariadb-account-delete" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.820231 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3895b57f-2dad-4d55-8818-e0c96b975933" containerName="mariadb-account-delete" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.820359 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-log" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.820377 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b09b59ae-48b2-4fda-80f3-591c6ce27dcd" containerName="glance-httpd" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.820383 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3895b57f-2dad-4d55-8818-e0c96b975933" containerName="mariadb-account-delete" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.820796 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.830726 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-1a26-account-create-update-ltfrh"] Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.832306 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.835517 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.835908 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-q9d6s"] Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.842649 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-1a26-account-create-update-ltfrh"] Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.917926 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bdc87e-a63e-4edb-a63b-af61c7175160-operator-scripts\") pod \"glance-db-create-q9d6s\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.918015 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbtjr\" (UniqueName: \"kubernetes.io/projected/5f68290c-e2cc-44c9-a321-afd6fd524542-kube-api-access-hbtjr\") pod \"glance-1a26-account-create-update-ltfrh\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.918258 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrtm4\" (UniqueName: \"kubernetes.io/projected/d2bdc87e-a63e-4edb-a63b-af61c7175160-kube-api-access-qrtm4\") pod \"glance-db-create-q9d6s\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:22 crc kubenswrapper[4706]: I1127 07:27:22.918373 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f68290c-e2cc-44c9-a321-afd6fd524542-operator-scripts\") pod \"glance-1a26-account-create-update-ltfrh\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.018978 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f68290c-e2cc-44c9-a321-afd6fd524542-operator-scripts\") pod \"glance-1a26-account-create-update-ltfrh\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.019039 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bdc87e-a63e-4edb-a63b-af61c7175160-operator-scripts\") pod \"glance-db-create-q9d6s\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.019095 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbtjr\" (UniqueName: \"kubernetes.io/projected/5f68290c-e2cc-44c9-a321-afd6fd524542-kube-api-access-hbtjr\") pod \"glance-1a26-account-create-update-ltfrh\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.019188 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrtm4\" (UniqueName: \"kubernetes.io/projected/d2bdc87e-a63e-4edb-a63b-af61c7175160-kube-api-access-qrtm4\") pod \"glance-db-create-q9d6s\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.021655 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bdc87e-a63e-4edb-a63b-af61c7175160-operator-scripts\") pod \"glance-db-create-q9d6s\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.021719 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f68290c-e2cc-44c9-a321-afd6fd524542-operator-scripts\") pod \"glance-1a26-account-create-update-ltfrh\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.036618 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrtm4\" (UniqueName: \"kubernetes.io/projected/d2bdc87e-a63e-4edb-a63b-af61c7175160-kube-api-access-qrtm4\") pod \"glance-db-create-q9d6s\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.040567 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbtjr\" (UniqueName: \"kubernetes.io/projected/5f68290c-e2cc-44c9-a321-afd6fd524542-kube-api-access-hbtjr\") pod \"glance-1a26-account-create-update-ltfrh\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.161316 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.166017 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.460725 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-1a26-account-create-update-ltfrh"] Nov 27 07:27:23 crc kubenswrapper[4706]: W1127 07:27:23.463413 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f68290c_e2cc_44c9_a321_afd6fd524542.slice/crio-1639f0abb7d8a37d48f30ed3a2089e8c91cc95f5b31d5d59ca1aef2f36e989b7 WatchSource:0}: Error finding container 1639f0abb7d8a37d48f30ed3a2089e8c91cc95f5b31d5d59ca1aef2f36e989b7: Status 404 returned error can't find the container with id 1639f0abb7d8a37d48f30ed3a2089e8c91cc95f5b31d5d59ca1aef2f36e989b7 Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.597989 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-q9d6s"] Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.658948 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" event={"ID":"5f68290c-e2cc-44c9-a321-afd6fd524542","Type":"ContainerStarted","Data":"b8bddeef39708a438e79d5792543906f80ab1d0ef1cf6d97d6def7030e1fddc5"} Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.659018 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" event={"ID":"5f68290c-e2cc-44c9-a321-afd6fd524542","Type":"ContainerStarted","Data":"1639f0abb7d8a37d48f30ed3a2089e8c91cc95f5b31d5d59ca1aef2f36e989b7"} Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.661107 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-q9d6s" event={"ID":"d2bdc87e-a63e-4edb-a63b-af61c7175160","Type":"ContainerStarted","Data":"cc8175fc3dde23e8022b6b0c1e50c62840f5ea24dc3191a900fd0010c518638d"} Nov 27 07:27:23 crc kubenswrapper[4706]: I1127 07:27:23.675431 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" podStartSLOduration=1.675413624 podStartE2EDuration="1.675413624s" podCreationTimestamp="2025-11-27 07:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:23.671329472 +0000 UTC m=+1127.560920282" watchObservedRunningTime="2025-11-27 07:27:23.675413624 +0000 UTC m=+1127.565004434" Nov 27 07:27:24 crc kubenswrapper[4706]: I1127 07:27:24.679261 4706 generic.go:334] "Generic (PLEG): container finished" podID="d2bdc87e-a63e-4edb-a63b-af61c7175160" containerID="d3830433d5c8e272ec09cf93094e448bf53b16a4d79969c062f25f5940a4e0f6" exitCode=0 Nov 27 07:27:24 crc kubenswrapper[4706]: I1127 07:27:24.679435 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-q9d6s" event={"ID":"d2bdc87e-a63e-4edb-a63b-af61c7175160","Type":"ContainerDied","Data":"d3830433d5c8e272ec09cf93094e448bf53b16a4d79969c062f25f5940a4e0f6"} Nov 27 07:27:24 crc kubenswrapper[4706]: I1127 07:27:24.685037 4706 generic.go:334] "Generic (PLEG): container finished" podID="5f68290c-e2cc-44c9-a321-afd6fd524542" containerID="b8bddeef39708a438e79d5792543906f80ab1d0ef1cf6d97d6def7030e1fddc5" exitCode=0 Nov 27 07:27:24 crc kubenswrapper[4706]: I1127 07:27:24.685114 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" event={"ID":"5f68290c-e2cc-44c9-a321-afd6fd524542","Type":"ContainerDied","Data":"b8bddeef39708a438e79d5792543906f80ab1d0ef1cf6d97d6def7030e1fddc5"} Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.171637 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.226654 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.277031 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrtm4\" (UniqueName: \"kubernetes.io/projected/d2bdc87e-a63e-4edb-a63b-af61c7175160-kube-api-access-qrtm4\") pod \"d2bdc87e-a63e-4edb-a63b-af61c7175160\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.277263 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bdc87e-a63e-4edb-a63b-af61c7175160-operator-scripts\") pod \"d2bdc87e-a63e-4edb-a63b-af61c7175160\" (UID: \"d2bdc87e-a63e-4edb-a63b-af61c7175160\") " Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.277298 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbtjr\" (UniqueName: \"kubernetes.io/projected/5f68290c-e2cc-44c9-a321-afd6fd524542-kube-api-access-hbtjr\") pod \"5f68290c-e2cc-44c9-a321-afd6fd524542\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.277318 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f68290c-e2cc-44c9-a321-afd6fd524542-operator-scripts\") pod \"5f68290c-e2cc-44c9-a321-afd6fd524542\" (UID: \"5f68290c-e2cc-44c9-a321-afd6fd524542\") " Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.278995 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2bdc87e-a63e-4edb-a63b-af61c7175160-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2bdc87e-a63e-4edb-a63b-af61c7175160" (UID: "d2bdc87e-a63e-4edb-a63b-af61c7175160"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.280347 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f68290c-e2cc-44c9-a321-afd6fd524542-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f68290c-e2cc-44c9-a321-afd6fd524542" (UID: "5f68290c-e2cc-44c9-a321-afd6fd524542"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.284348 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f68290c-e2cc-44c9-a321-afd6fd524542-kube-api-access-hbtjr" (OuterVolumeSpecName: "kube-api-access-hbtjr") pod "5f68290c-e2cc-44c9-a321-afd6fd524542" (UID: "5f68290c-e2cc-44c9-a321-afd6fd524542"). InnerVolumeSpecName "kube-api-access-hbtjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.284365 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2bdc87e-a63e-4edb-a63b-af61c7175160-kube-api-access-qrtm4" (OuterVolumeSpecName: "kube-api-access-qrtm4") pod "d2bdc87e-a63e-4edb-a63b-af61c7175160" (UID: "d2bdc87e-a63e-4edb-a63b-af61c7175160"). InnerVolumeSpecName "kube-api-access-qrtm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.379001 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bdc87e-a63e-4edb-a63b-af61c7175160-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.379039 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbtjr\" (UniqueName: \"kubernetes.io/projected/5f68290c-e2cc-44c9-a321-afd6fd524542-kube-api-access-hbtjr\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.379054 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f68290c-e2cc-44c9-a321-afd6fd524542-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.379066 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrtm4\" (UniqueName: \"kubernetes.io/projected/d2bdc87e-a63e-4edb-a63b-af61c7175160-kube-api-access-qrtm4\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.705163 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-q9d6s" event={"ID":"d2bdc87e-a63e-4edb-a63b-af61c7175160","Type":"ContainerDied","Data":"cc8175fc3dde23e8022b6b0c1e50c62840f5ea24dc3191a900fd0010c518638d"} Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.705207 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc8175fc3dde23e8022b6b0c1e50c62840f5ea24dc3191a900fd0010c518638d" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.705212 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-q9d6s" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.707110 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" event={"ID":"5f68290c-e2cc-44c9-a321-afd6fd524542","Type":"ContainerDied","Data":"1639f0abb7d8a37d48f30ed3a2089e8c91cc95f5b31d5d59ca1aef2f36e989b7"} Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.707136 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1639f0abb7d8a37d48f30ed3a2089e8c91cc95f5b31d5d59ca1aef2f36e989b7" Nov 27 07:27:26 crc kubenswrapper[4706]: I1127 07:27:26.707194 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-1a26-account-create-update-ltfrh" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.040155 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-6lghz"] Nov 27 07:27:28 crc kubenswrapper[4706]: E1127 07:27:28.041112 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f68290c-e2cc-44c9-a321-afd6fd524542" containerName="mariadb-account-create-update" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.041148 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f68290c-e2cc-44c9-a321-afd6fd524542" containerName="mariadb-account-create-update" Nov 27 07:27:28 crc kubenswrapper[4706]: E1127 07:27:28.041194 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2bdc87e-a63e-4edb-a63b-af61c7175160" containerName="mariadb-database-create" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.041210 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2bdc87e-a63e-4edb-a63b-af61c7175160" containerName="mariadb-database-create" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.041469 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f68290c-e2cc-44c9-a321-afd6fd524542" containerName="mariadb-account-create-update" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.041504 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2bdc87e-a63e-4edb-a63b-af61c7175160" containerName="mariadb-database-create" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.042257 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.048950 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-q9gnm" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.059449 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.063321 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6lghz"] Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.104090 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-config-data\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.104161 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-db-sync-config-data\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.104192 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5trf5\" (UniqueName: \"kubernetes.io/projected/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-kube-api-access-5trf5\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.205065 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-config-data\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.205127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-db-sync-config-data\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.205158 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5trf5\" (UniqueName: \"kubernetes.io/projected/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-kube-api-access-5trf5\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.220259 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-db-sync-config-data\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.221808 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-config-data\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.234966 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5trf5\" (UniqueName: \"kubernetes.io/projected/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-kube-api-access-5trf5\") pod \"glance-db-sync-6lghz\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.371556 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.682091 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6lghz"] Nov 27 07:27:28 crc kubenswrapper[4706]: I1127 07:27:28.724016 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6lghz" event={"ID":"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608","Type":"ContainerStarted","Data":"abbcf19ca4b57d44b4b572579d9a9584fb7fe9095df8e85897042693c108c565"} Nov 27 07:27:29 crc kubenswrapper[4706]: I1127 07:27:29.732509 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6lghz" event={"ID":"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608","Type":"ContainerStarted","Data":"233302eba3934529a083f1c8281a7359f350ca8cfdb0961953fe75c70dbf51ae"} Nov 27 07:27:29 crc kubenswrapper[4706]: I1127 07:27:29.750686 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-6lghz" podStartSLOduration=1.750667339 podStartE2EDuration="1.750667339s" podCreationTimestamp="2025-11-27 07:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:29.749571943 +0000 UTC m=+1133.639162763" watchObservedRunningTime="2025-11-27 07:27:29.750667339 +0000 UTC m=+1133.640258149" Nov 27 07:27:32 crc kubenswrapper[4706]: I1127 07:27:32.763472 4706 generic.go:334] "Generic (PLEG): container finished" podID="b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" containerID="233302eba3934529a083f1c8281a7359f350ca8cfdb0961953fe75c70dbf51ae" exitCode=0 Nov 27 07:27:32 crc kubenswrapper[4706]: I1127 07:27:32.763575 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6lghz" event={"ID":"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608","Type":"ContainerDied","Data":"233302eba3934529a083f1c8281a7359f350ca8cfdb0961953fe75c70dbf51ae"} Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.184844 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.195967 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5trf5\" (UniqueName: \"kubernetes.io/projected/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-kube-api-access-5trf5\") pod \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.196070 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-config-data\") pod \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.196093 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-db-sync-config-data\") pod \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\" (UID: \"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608\") " Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.204141 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-kube-api-access-5trf5" (OuterVolumeSpecName: "kube-api-access-5trf5") pod "b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" (UID: "b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608"). InnerVolumeSpecName "kube-api-access-5trf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.215622 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" (UID: "b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.289930 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-config-data" (OuterVolumeSpecName: "config-data") pod "b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" (UID: "b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.297690 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5trf5\" (UniqueName: \"kubernetes.io/projected/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-kube-api-access-5trf5\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.297718 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.297728 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.790192 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6lghz" Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.797214 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6lghz" event={"ID":"b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608","Type":"ContainerDied","Data":"abbcf19ca4b57d44b4b572579d9a9584fb7fe9095df8e85897042693c108c565"} Nov 27 07:27:34 crc kubenswrapper[4706]: I1127 07:27:34.797295 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abbcf19ca4b57d44b4b572579d9a9584fb7fe9095df8e85897042693c108c565" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.185009 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:27:36 crc kubenswrapper[4706]: E1127 07:27:36.185285 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" containerName="glance-db-sync" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.185297 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" containerName="glance-db-sync" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.185429 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" containerName="glance-db-sync" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.186275 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.188408 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.189152 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-q9gnm" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.189332 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.209000 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.336632 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.336728 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-config-data\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.336780 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.336799 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.336816 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-scripts\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.336859 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg9dv\" (UniqueName: \"kubernetes.io/projected/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-kube-api-access-jg9dv\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.336975 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-logs\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.337025 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.337045 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-run\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.337112 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.337178 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.337203 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-sys\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.337247 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.337432 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-dev\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.438866 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.438913 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.438939 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-sys\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.438956 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.438987 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-dev\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439007 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439027 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-config-data\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439051 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439065 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439078 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-scripts\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439131 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg9dv\" (UniqueName: \"kubernetes.io/projected/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-kube-api-access-jg9dv\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439179 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-logs\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439211 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439244 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-run\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439307 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-run\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439600 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.439962 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440173 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440273 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440308 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-sys\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440370 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-logs\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440417 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440442 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-dev\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440487 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.440672 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.452121 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-scripts\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.452653 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-config-data\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.458816 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.459789 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg9dv\" (UniqueName: \"kubernetes.io/projected/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-kube-api-access-jg9dv\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.462797 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.497452 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.498422 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.498797 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.516578 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.522908 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.642892 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-logs\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643146 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-sys\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643212 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643246 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-dev\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643270 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643289 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-run\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643302 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643317 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643331 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643346 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643374 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643397 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643413 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxv4k\" (UniqueName: \"kubernetes.io/projected/0cc04539-210c-4048-bd89-190e2e24632e-kube-api-access-gxv4k\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.643437 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744706 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744777 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-logs\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744799 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-sys\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744832 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744850 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-dev\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744887 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744903 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-run\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744919 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744935 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744949 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744975 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.744996 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.745014 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxv4k\" (UniqueName: \"kubernetes.io/projected/0cc04539-210c-4048-bd89-190e2e24632e-kube-api-access-gxv4k\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.745375 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.745810 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-logs\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.745848 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-sys\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.746163 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.746841 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.746910 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-dev\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.746938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.747615 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-run\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.747650 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.748093 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.748412 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.749474 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.760244 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.763764 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.770961 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.776035 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxv4k\" (UniqueName: \"kubernetes.io/projected/0cc04539-210c-4048-bd89-190e2e24632e-kube-api-access-gxv4k\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.787772 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.833029 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:36 crc kubenswrapper[4706]: I1127 07:27:36.984740 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.058831 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.202611 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.813266 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerStarted","Data":"efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.814056 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerStarted","Data":"25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.814116 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerStarted","Data":"66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.814177 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerStarted","Data":"f5fcaa586207bc23d8cee7b0d97337dd74c758c1e0200c0ad19629492cf3e795"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.816316 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerStarted","Data":"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.816369 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerStarted","Data":"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.816390 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerStarted","Data":"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.816408 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerStarted","Data":"55b1297cc4e51c465e01ef6121a06b77d8ec95f61454f1034eec58ddba95027e"} Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.816492 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-log" containerID="cri-o://f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39" gracePeriod=30 Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.816589 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-api" containerID="cri-o://a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674" gracePeriod=30 Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.816611 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-httpd" containerID="cri-o://683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a" gracePeriod=30 Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.860487 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=1.860463886 podStartE2EDuration="1.860463886s" podCreationTimestamp="2025-11-27 07:27:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:37.847286873 +0000 UTC m=+1141.736877683" watchObservedRunningTime="2025-11-27 07:27:37.860463886 +0000 UTC m=+1141.750054706" Nov 27 07:27:37 crc kubenswrapper[4706]: I1127 07:27:37.893867 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.893837282 podStartE2EDuration="2.893837282s" podCreationTimestamp="2025-11-27 07:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:37.883189328 +0000 UTC m=+1141.772780168" watchObservedRunningTime="2025-11-27 07:27:37.893837282 +0000 UTC m=+1141.783428102" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.337414 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388433 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-logs\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388495 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-nvme\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388522 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388544 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388594 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-scripts\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388628 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388654 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-lib-modules\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388697 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388736 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-run\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388776 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-sys\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388817 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-run" (OuterVolumeSpecName: "run") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388840 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-httpd-run\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388858 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-sys" (OuterVolumeSpecName: "sys") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388870 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-dev\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388902 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-var-locks-brick\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388898 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-logs" (OuterVolumeSpecName: "logs") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388928 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-iscsi\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.388962 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389001 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-config-data\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389003 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-dev" (OuterVolumeSpecName: "dev") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389056 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389087 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxv4k\" (UniqueName: \"kubernetes.io/projected/0cc04539-210c-4048-bd89-190e2e24632e-kube-api-access-gxv4k\") pod \"0cc04539-210c-4048-bd89-190e2e24632e\" (UID: \"0cc04539-210c-4048-bd89-190e2e24632e\") " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389109 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389619 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389635 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389648 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389659 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389669 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389679 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cc04539-210c-4048-bd89-190e2e24632e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389690 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389701 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.389712 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0cc04539-210c-4048-bd89-190e2e24632e-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.395703 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.396271 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc04539-210c-4048-bd89-190e2e24632e-kube-api-access-gxv4k" (OuterVolumeSpecName: "kube-api-access-gxv4k") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "kube-api-access-gxv4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.398430 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.399995 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-scripts" (OuterVolumeSpecName: "scripts") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.479864 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-config-data" (OuterVolumeSpecName: "config-data") pod "0cc04539-210c-4048-bd89-190e2e24632e" (UID: "0cc04539-210c-4048-bd89-190e2e24632e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.490781 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxv4k\" (UniqueName: \"kubernetes.io/projected/0cc04539-210c-4048-bd89-190e2e24632e-kube-api-access-gxv4k\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.490838 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.490868 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.490880 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.490890 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc04539-210c-4048-bd89-190e2e24632e-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.504242 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.508604 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.592859 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.593109 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.827963 4706 generic.go:334] "Generic (PLEG): container finished" podID="0cc04539-210c-4048-bd89-190e2e24632e" containerID="a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674" exitCode=143 Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.828005 4706 generic.go:334] "Generic (PLEG): container finished" podID="0cc04539-210c-4048-bd89-190e2e24632e" containerID="683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a" exitCode=143 Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.828018 4706 generic.go:334] "Generic (PLEG): container finished" podID="0cc04539-210c-4048-bd89-190e2e24632e" containerID="f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39" exitCode=143 Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.828047 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.828461 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerDied","Data":"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674"} Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.828722 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerDied","Data":"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a"} Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.828895 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerDied","Data":"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39"} Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.829050 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0cc04539-210c-4048-bd89-190e2e24632e","Type":"ContainerDied","Data":"55b1297cc4e51c465e01ef6121a06b77d8ec95f61454f1034eec58ddba95027e"} Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.829201 4706 scope.go:117] "RemoveContainer" containerID="a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.854536 4706 scope.go:117] "RemoveContainer" containerID="683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.855453 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.873546 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.882771 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:38 crc kubenswrapper[4706]: E1127 07:27:38.883104 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-api" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.883126 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-api" Nov 27 07:27:38 crc kubenswrapper[4706]: E1127 07:27:38.883146 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-log" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.883153 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-log" Nov 27 07:27:38 crc kubenswrapper[4706]: E1127 07:27:38.883165 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-httpd" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.883172 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-httpd" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.883344 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-log" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.883358 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-api" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.883368 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc04539-210c-4048-bd89-190e2e24632e" containerName="glance-httpd" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.884104 4706 scope.go:117] "RemoveContainer" containerID="f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.884440 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.886987 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.906098 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.929643 4706 scope.go:117] "RemoveContainer" containerID="a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674" Nov 27 07:27:38 crc kubenswrapper[4706]: E1127 07:27:38.930016 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674\": container with ID starting with a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674 not found: ID does not exist" containerID="a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.930056 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674"} err="failed to get container status \"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674\": rpc error: code = NotFound desc = could not find container \"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674\": container with ID starting with a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674 not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.930081 4706 scope.go:117] "RemoveContainer" containerID="683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a" Nov 27 07:27:38 crc kubenswrapper[4706]: E1127 07:27:38.930439 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a\": container with ID starting with 683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a not found: ID does not exist" containerID="683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.930511 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a"} err="failed to get container status \"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a\": rpc error: code = NotFound desc = could not find container \"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a\": container with ID starting with 683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.930536 4706 scope.go:117] "RemoveContainer" containerID="f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39" Nov 27 07:27:38 crc kubenswrapper[4706]: E1127 07:27:38.930828 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39\": container with ID starting with f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39 not found: ID does not exist" containerID="f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.930893 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39"} err="failed to get container status \"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39\": rpc error: code = NotFound desc = could not find container \"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39\": container with ID starting with f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39 not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.930915 4706 scope.go:117] "RemoveContainer" containerID="a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.931158 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674"} err="failed to get container status \"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674\": rpc error: code = NotFound desc = could not find container \"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674\": container with ID starting with a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674 not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.931179 4706 scope.go:117] "RemoveContainer" containerID="683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.931440 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a"} err="failed to get container status \"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a\": rpc error: code = NotFound desc = could not find container \"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a\": container with ID starting with 683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.931465 4706 scope.go:117] "RemoveContainer" containerID="f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.933492 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39"} err="failed to get container status \"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39\": rpc error: code = NotFound desc = could not find container \"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39\": container with ID starting with f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39 not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.933515 4706 scope.go:117] "RemoveContainer" containerID="a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.933695 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674"} err="failed to get container status \"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674\": rpc error: code = NotFound desc = could not find container \"a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674\": container with ID starting with a3b79bd26fb7d902c82502b4f490c41cc42dc74f3c9460486ff7099fd2bdc674 not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.933712 4706 scope.go:117] "RemoveContainer" containerID="683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.934316 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a"} err="failed to get container status \"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a\": rpc error: code = NotFound desc = could not find container \"683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a\": container with ID starting with 683834ffecd0e25057a855bfa70571b6985c127701a40ebf3e5e53f20b23ef2a not found: ID does not exist" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.934338 4706 scope.go:117] "RemoveContainer" containerID="f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39" Nov 27 07:27:38 crc kubenswrapper[4706]: I1127 07:27:38.934711 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39"} err="failed to get container status \"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39\": rpc error: code = NotFound desc = could not find container \"f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39\": container with ID starting with f78ef695b3690e2714645d048105afa23abe80377767dd9d5c2f7d59ce014c39 not found: ID does not exist" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002306 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-run\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002386 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-dev\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002402 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002421 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002537 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002624 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002665 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002695 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-sys\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002732 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002814 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002843 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002863 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.002989 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97tdk\" (UniqueName: \"kubernetes.io/projected/3c6cf53c-105b-4caa-9c94-f7363752c64a-kube-api-access-97tdk\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.003094 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104702 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104770 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104804 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-sys\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104827 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104852 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104875 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104887 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104902 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.104993 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97tdk\" (UniqueName: \"kubernetes.io/projected/3c6cf53c-105b-4caa-9c94-f7363752c64a-kube-api-access-97tdk\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105053 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105094 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-run\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105133 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-dev\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105150 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105301 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105407 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105461 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105494 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-sys\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.105723 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.106435 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-run\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.106495 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.106534 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-dev\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.106506 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.106586 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.107132 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.125515 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.135115 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.138637 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97tdk\" (UniqueName: \"kubernetes.io/projected/3c6cf53c-105b-4caa-9c94-f7363752c64a-kube-api-access-97tdk\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.153892 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.168809 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.211911 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.697023 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:27:39 crc kubenswrapper[4706]: W1127 07:27:39.715969 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c6cf53c_105b_4caa_9c94_f7363752c64a.slice/crio-ed9ba87fcb42e2e2a2a4a7ef195b329f94879dd10fc232dd103bfdd3284d4f8f WatchSource:0}: Error finding container ed9ba87fcb42e2e2a2a4a7ef195b329f94879dd10fc232dd103bfdd3284d4f8f: Status 404 returned error can't find the container with id ed9ba87fcb42e2e2a2a4a7ef195b329f94879dd10fc232dd103bfdd3284d4f8f Nov 27 07:27:39 crc kubenswrapper[4706]: I1127 07:27:39.842155 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerStarted","Data":"ed9ba87fcb42e2e2a2a4a7ef195b329f94879dd10fc232dd103bfdd3284d4f8f"} Nov 27 07:27:40 crc kubenswrapper[4706]: I1127 07:27:40.798753 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cc04539-210c-4048-bd89-190e2e24632e" path="/var/lib/kubelet/pods/0cc04539-210c-4048-bd89-190e2e24632e/volumes" Nov 27 07:27:40 crc kubenswrapper[4706]: I1127 07:27:40.857478 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerStarted","Data":"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188"} Nov 27 07:27:40 crc kubenswrapper[4706]: I1127 07:27:40.857851 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerStarted","Data":"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa"} Nov 27 07:27:40 crc kubenswrapper[4706]: I1127 07:27:40.858120 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerStarted","Data":"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a"} Nov 27 07:27:40 crc kubenswrapper[4706]: I1127 07:27:40.906575 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.906540296 podStartE2EDuration="2.906540296s" podCreationTimestamp="2025-11-27 07:27:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:40.897409239 +0000 UTC m=+1144.787000079" watchObservedRunningTime="2025-11-27 07:27:40.906540296 +0000 UTC m=+1144.796131136" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.499694 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.500265 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.500286 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.534920 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.536570 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.537362 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.919933 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.919971 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.919980 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.930384 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.930659 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:46 crc kubenswrapper[4706]: I1127 07:27:46.930777 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.213392 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.213652 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.213662 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.249594 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.254985 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.278279 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.944509 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.944547 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.944559 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.962217 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.964064 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:49 crc kubenswrapper[4706]: I1127 07:27:49.965638 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.132465 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.135673 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.147677 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.150196 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.159560 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.172160 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.245565 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.247606 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.269483 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.270962 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.287287 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.302533 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.348930 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-sys\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.348980 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-config-data\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349000 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349021 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349038 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-dev\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349051 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-logs\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349072 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349091 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349156 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349183 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-config-data\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349204 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-run\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349253 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349270 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-scripts\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349289 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkrhz\" (UniqueName: \"kubernetes.io/projected/3e01c94a-8908-4a33-a45a-41f48b44217b-kube-api-access-mkrhz\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349331 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-logs\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349345 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-scripts\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349361 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-sys\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349377 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5qgd\" (UniqueName: \"kubernetes.io/projected/ab67a0f0-0b9e-436e-b62b-07b5f34de477-kube-api-access-n5qgd\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349400 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349419 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349434 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-dev\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349451 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349464 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-run\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349484 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349505 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349528 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.349546 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450293 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450327 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-logs\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450346 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-scripts\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450365 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-sys\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450383 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5qgd\" (UniqueName: \"kubernetes.io/projected/ab67a0f0-0b9e-436e-b62b-07b5f34de477-kube-api-access-n5qgd\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450411 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-config-data\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450427 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-logs\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450446 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450461 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-dev\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450475 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450495 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450516 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450533 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450550 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-dev\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450582 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-run\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450604 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450627 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450645 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450660 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450674 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450683 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450732 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450766 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-sys\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451018 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-logs\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451032 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451058 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451088 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-dev\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451112 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451132 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-run\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451250 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.450694 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451506 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451531 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-run\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451553 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451580 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451602 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-sys\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451625 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-config-data\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451646 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451671 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451698 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451717 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-dev\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451731 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-logs\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451753 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451778 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451800 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451820 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451845 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451860 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb8d9\" (UniqueName: \"kubernetes.io/projected/d4312c89-9114-487a-807c-3d5510238c4d-kube-api-access-pb8d9\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451886 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-sys\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451913 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451935 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-logs\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-sys\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.451989 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-config-data\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452015 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-run\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452033 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-dev\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452049 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452062 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-scripts\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452086 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-scripts\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452105 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-config-data\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452132 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-run\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452153 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6v25\" (UniqueName: \"kubernetes.io/projected/0b2dd611-2516-4714-8a41-c7ebc3fcc045-kube-api-access-h6v25\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452174 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452195 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-scripts\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452214 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkrhz\" (UniqueName: \"kubernetes.io/projected/3e01c94a-8908-4a33-a45a-41f48b44217b-kube-api-access-mkrhz\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452247 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452393 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.452467 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454497 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-sys\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454597 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454592 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-run\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454708 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454740 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454789 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454760 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454836 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-logs\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.454897 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-dev\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.455011 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.457707 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-scripts\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.462751 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-scripts\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.463281 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-config-data\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.465388 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-config-data\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.467404 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5qgd\" (UniqueName: \"kubernetes.io/projected/ab67a0f0-0b9e-436e-b62b-07b5f34de477-kube-api-access-n5qgd\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.488886 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkrhz\" (UniqueName: \"kubernetes.io/projected/3e01c94a-8908-4a33-a45a-41f48b44217b-kube-api-access-mkrhz\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.489633 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.491659 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.495653 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.515938 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.553800 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-config-data\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.553871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-logs\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.553909 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.553943 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-dev\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.553990 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554100 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554123 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554146 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554190 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554217 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554269 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554311 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-run\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554345 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554380 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554423 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554442 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554453 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554459 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-logs\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554562 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554681 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554686 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554705 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-run\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554754 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554797 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554821 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554874 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554884 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-dev\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554918 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554945 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.554961 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb8d9\" (UniqueName: \"kubernetes.io/projected/d4312c89-9114-487a-807c-3d5510238c4d-kube-api-access-pb8d9\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555012 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-sys\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555047 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-logs\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555054 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555083 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-sys\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555131 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-scripts\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555163 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-dev\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555194 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555260 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-scripts\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555289 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-config-data\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555318 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-dev\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555331 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-run\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555344 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-sys\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555093 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-sys\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555364 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6v25\" (UniqueName: \"kubernetes.io/projected/0b2dd611-2516-4714-8a41-c7ebc3fcc045-kube-api-access-h6v25\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555298 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-logs\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555403 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555446 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555486 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-run\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.555708 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.556189 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.557487 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-config-data\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.561886 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-scripts\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.562107 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-config-data\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.568449 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-scripts\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.580580 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6v25\" (UniqueName: \"kubernetes.io/projected/0b2dd611-2516-4714-8a41-c7ebc3fcc045-kube-api-access-h6v25\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.581690 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb8d9\" (UniqueName: \"kubernetes.io/projected/d4312c89-9114-487a-807c-3d5510238c4d-kube-api-access-pb8d9\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.584821 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.588058 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.594888 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.600410 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.605954 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.771561 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.792751 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.801326 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:27:52 crc kubenswrapper[4706]: W1127 07:27:52.816682 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b2dd611_2516_4714_8a41_c7ebc3fcc045.slice/crio-eb5afdc99442a93c93b4d53d615821a4c748c6baac6bcaab0ebfd05638ae723c WatchSource:0}: Error finding container eb5afdc99442a93c93b4d53d615821a4c748c6baac6bcaab0ebfd05638ae723c: Status 404 returned error can't find the container with id eb5afdc99442a93c93b4d53d615821a4c748c6baac6bcaab0ebfd05638ae723c Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.873613 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:27:52 crc kubenswrapper[4706]: I1127 07:27:52.974130 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerStarted","Data":"eb5afdc99442a93c93b4d53d615821a4c748c6baac6bcaab0ebfd05638ae723c"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.048229 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:27:53 crc kubenswrapper[4706]: W1127 07:27:53.064159 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e01c94a_8908_4a33_a45a_41f48b44217b.slice/crio-7ef0c86d56b6bd9ef153ebfa4176fc93f19ba9f13a3aa1b1dcdf85dade3ab0e8 WatchSource:0}: Error finding container 7ef0c86d56b6bd9ef153ebfa4176fc93f19ba9f13a3aa1b1dcdf85dade3ab0e8: Status 404 returned error can't find the container with id 7ef0c86d56b6bd9ef153ebfa4176fc93f19ba9f13a3aa1b1dcdf85dade3ab0e8 Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.339376 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.393473 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:27:53 crc kubenswrapper[4706]: W1127 07:27:53.416447 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4312c89_9114_487a_807c_3d5510238c4d.slice/crio-1fa8f953aba698308ce9037df202b922f539e70bf2b423270196c685a30b65fa WatchSource:0}: Error finding container 1fa8f953aba698308ce9037df202b922f539e70bf2b423270196c685a30b65fa: Status 404 returned error can't find the container with id 1fa8f953aba698308ce9037df202b922f539e70bf2b423270196c685a30b65fa Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.984708 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerStarted","Data":"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.985478 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerStarted","Data":"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.985497 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerStarted","Data":"1fa8f953aba698308ce9037df202b922f539e70bf2b423270196c685a30b65fa"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.987002 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerStarted","Data":"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.987064 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerStarted","Data":"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.987081 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerStarted","Data":"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.987096 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerStarted","Data":"7ef0c86d56b6bd9ef153ebfa4176fc93f19ba9f13a3aa1b1dcdf85dade3ab0e8"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.990770 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerStarted","Data":"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.990949 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerStarted","Data":"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.991128 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerStarted","Data":"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.997325 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerStarted","Data":"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.997367 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerStarted","Data":"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.997383 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerStarted","Data":"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413"} Nov 27 07:27:53 crc kubenswrapper[4706]: I1127 07:27:53.997395 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerStarted","Data":"55d791d9a3ef38ddbfe4ac6a2d4f4c542c2028be50f2172ca85c2f647373fc87"} Nov 27 07:27:54 crc kubenswrapper[4706]: I1127 07:27:54.019182 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.019163002 podStartE2EDuration="3.019163002s" podCreationTimestamp="2025-11-27 07:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:54.014425344 +0000 UTC m=+1157.904016184" watchObservedRunningTime="2025-11-27 07:27:54.019163002 +0000 UTC m=+1157.908753812" Nov 27 07:27:54 crc kubenswrapper[4706]: I1127 07:27:54.061158 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.061132848 podStartE2EDuration="3.061132848s" podCreationTimestamp="2025-11-27 07:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:54.051579909 +0000 UTC m=+1157.941170709" watchObservedRunningTime="2025-11-27 07:27:54.061132848 +0000 UTC m=+1157.950723678" Nov 27 07:27:54 crc kubenswrapper[4706]: I1127 07:27:54.090619 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.090600463 podStartE2EDuration="3.090600463s" podCreationTimestamp="2025-11-27 07:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:54.08315426 +0000 UTC m=+1157.972745160" watchObservedRunningTime="2025-11-27 07:27:54.090600463 +0000 UTC m=+1157.980191273" Nov 27 07:27:55 crc kubenswrapper[4706]: I1127 07:27:55.008233 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerStarted","Data":"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751"} Nov 27 07:27:55 crc kubenswrapper[4706]: I1127 07:27:55.058351 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=4.058331973 podStartE2EDuration="4.058331973s" podCreationTimestamp="2025-11-27 07:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:27:55.051351744 +0000 UTC m=+1158.940942554" watchObservedRunningTime="2025-11-27 07:27:55.058331973 +0000 UTC m=+1158.947922783" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.600759 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.601450 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.601469 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.650872 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.656614 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.683710 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.772397 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.772647 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.772727 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.795770 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.795814 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.796447 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.810185 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.821012 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.826951 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.827333 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.828725 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.837543 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.874560 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.875723 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.875747 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.904947 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.911775 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:02 crc kubenswrapper[4706]: I1127 07:28:02.922397 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095337 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095716 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095737 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095745 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095756 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095766 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095774 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.095782 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.096680 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.096714 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.096726 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.096736 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.111827 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.113736 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.116269 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.116751 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.117603 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.118456 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.120048 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.121924 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.122476 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.123689 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.125523 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.132373 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.838167 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:28:03 crc kubenswrapper[4706]: I1127 07:28:03.844415 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:28:04 crc kubenswrapper[4706]: I1127 07:28:04.024518 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:28:04 crc kubenswrapper[4706]: I1127 07:28:04.033067 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.116347 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-log" containerID="cri-o://9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.116506 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-api" containerID="cri-o://746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.116578 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-httpd" containerID="cri-o://250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.116726 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-log" containerID="cri-o://5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.116790 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-httpd" containerID="cri-o://0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.116801 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-api" containerID="cri-o://59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.117527 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-log" containerID="cri-o://3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.117588 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-api" containerID="cri-o://82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.117665 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-httpd" containerID="cri-o://aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.117774 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-log" containerID="cri-o://f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.117794 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-api" containerID="cri-o://68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.117803 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-httpd" containerID="cri-o://b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121" gracePeriod=30 Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.962931 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:05 crc kubenswrapper[4706]: I1127 07:28:05.967863 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.021762 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-dev\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.021820 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-iscsi\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.021872 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkrhz\" (UniqueName: \"kubernetes.io/projected/3e01c94a-8908-4a33-a45a-41f48b44217b-kube-api-access-mkrhz\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.021903 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-config-data\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.021937 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-config-data\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.021979 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022007 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022037 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022058 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-sys\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022087 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-lib-modules\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022114 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022145 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-run\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022146 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.022202 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-dev" (OuterVolumeSpecName: "dev") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.023022 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.023087 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-sys" (OuterVolumeSpecName: "sys") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.024598 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-run" (OuterVolumeSpecName: "run") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.026433 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-httpd-run\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.026475 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-dev\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027079 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-httpd-run\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027109 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-logs\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027149 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-nvme\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027189 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-scripts\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027215 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-scripts\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027255 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6v25\" (UniqueName: \"kubernetes.io/projected/0b2dd611-2516-4714-8a41-c7ebc3fcc045-kube-api-access-h6v25\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027279 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-logs\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027301 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-var-locks-brick\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027328 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-iscsi\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027348 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-run\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027363 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-nvme\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-var-locks-brick\") pod \"3e01c94a-8908-4a33-a45a-41f48b44217b\" (UID: \"3e01c94a-8908-4a33-a45a-41f48b44217b\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027022 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-dev" (OuterVolumeSpecName: "dev") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027399 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-lib-modules\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027162 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027389 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027422 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-sys\") pod \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\" (UID: \"0b2dd611-2516-4714-8a41-c7ebc3fcc045\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027445 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-sys" (OuterVolumeSpecName: "sys") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.027595 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028016 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028034 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028045 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028053 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028061 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028068 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028076 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028084 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028091 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028117 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028126 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028154 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028174 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028192 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-run" (OuterVolumeSpecName: "run") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028872 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.028874 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.033416 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.033556 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-logs" (OuterVolumeSpecName: "logs") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.036185 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e01c94a-8908-4a33-a45a-41f48b44217b-kube-api-access-mkrhz" (OuterVolumeSpecName: "kube-api-access-mkrhz") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "kube-api-access-mkrhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.041822 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-logs" (OuterVolumeSpecName: "logs") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.047162 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.047302 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.048701 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2dd611-2516-4714-8a41-c7ebc3fcc045-kube-api-access-h6v25" (OuterVolumeSpecName: "kube-api-access-h6v25") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "kube-api-access-h6v25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.048861 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-scripts" (OuterVolumeSpecName: "scripts") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.052414 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance-cache") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.057332 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-scripts" (OuterVolumeSpecName: "scripts") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.072468 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.128727 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-logs\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.128784 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.128806 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-config-data\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.128864 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb8d9\" (UniqueName: \"kubernetes.io/projected/d4312c89-9114-487a-807c-3d5510238c4d-kube-api-access-pb8d9\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.128897 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-httpd-run\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.128938 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-lib-modules\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.128977 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-scripts\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129029 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-dev\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129048 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-var-locks-brick\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129084 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-nvme\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129109 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-iscsi\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129123 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-logs" (OuterVolumeSpecName: "logs") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129132 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129181 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-run\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129216 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-sys\") pod \"d4312c89-9114-487a-807c-3d5510238c4d\" (UID: \"d4312c89-9114-487a-807c-3d5510238c4d\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129766 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkrhz\" (UniqueName: \"kubernetes.io/projected/3e01c94a-8908-4a33-a45a-41f48b44217b-kube-api-access-mkrhz\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129960 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129973 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129986 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.129998 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130012 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130022 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e01c94a-8908-4a33-a45a-41f48b44217b-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130030 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130038 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130047 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6v25\" (UniqueName: \"kubernetes.io/projected/0b2dd611-2516-4714-8a41-c7ebc3fcc045-kube-api-access-h6v25\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130057 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2dd611-2516-4714-8a41-c7ebc3fcc045-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130066 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130074 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130083 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130092 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e01c94a-8908-4a33-a45a-41f48b44217b-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.130100 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0b2dd611-2516-4714-8a41-c7ebc3fcc045-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131463 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-run" (OuterVolumeSpecName: "run") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131524 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-sys" (OuterVolumeSpecName: "sys") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131552 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131571 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-dev" (OuterVolumeSpecName: "dev") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131838 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131882 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131906 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131919 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.131927 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.132185 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.133473 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135740 4706 generic.go:334] "Generic (PLEG): container finished" podID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerID="68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135768 4706 generic.go:334] "Generic (PLEG): container finished" podID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerID="b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135778 4706 generic.go:334] "Generic (PLEG): container finished" podID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerID="f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432" exitCode=143 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135867 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerDied","Data":"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135927 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerDied","Data":"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135948 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerDied","Data":"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135964 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0b2dd611-2516-4714-8a41-c7ebc3fcc045","Type":"ContainerDied","Data":"eb5afdc99442a93c93b4d53d615821a4c748c6baac6bcaab0ebfd05638ae723c"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.135981 4706 scope.go:117] "RemoveContainer" containerID="68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.136153 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.138846 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-scripts" (OuterVolumeSpecName: "scripts") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139260 4706 generic.go:334] "Generic (PLEG): container finished" podID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerID="746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139285 4706 generic.go:334] "Generic (PLEG): container finished" podID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerID="250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139294 4706 generic.go:334] "Generic (PLEG): container finished" podID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerID="9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413" exitCode=143 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139312 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139340 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerDied","Data":"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139365 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerDied","Data":"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139375 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerDied","Data":"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.139384 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ab67a0f0-0b9e-436e-b62b-07b5f34de477","Type":"ContainerDied","Data":"55d791d9a3ef38ddbfe4ac6a2d4f4c542c2028be50f2172ca85c2f647373fc87"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.141166 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-config-data" (OuterVolumeSpecName: "config-data") pod "3e01c94a-8908-4a33-a45a-41f48b44217b" (UID: "3e01c94a-8908-4a33-a45a-41f48b44217b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.141463 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4312c89-9114-487a-807c-3d5510238c4d-kube-api-access-pb8d9" (OuterVolumeSpecName: "kube-api-access-pb8d9") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "kube-api-access-pb8d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144384 4706 generic.go:334] "Generic (PLEG): container finished" podID="d4312c89-9114-487a-807c-3d5510238c4d" containerID="59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144411 4706 generic.go:334] "Generic (PLEG): container finished" podID="d4312c89-9114-487a-807c-3d5510238c4d" containerID="0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144421 4706 generic.go:334] "Generic (PLEG): container finished" podID="d4312c89-9114-487a-807c-3d5510238c4d" containerID="5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d" exitCode=143 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144463 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerDied","Data":"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144490 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerDied","Data":"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144500 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerDied","Data":"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144508 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"d4312c89-9114-487a-807c-3d5510238c4d","Type":"ContainerDied","Data":"1fa8f953aba698308ce9037df202b922f539e70bf2b423270196c685a30b65fa"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.144551 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.146401 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-config-data" (OuterVolumeSpecName: "config-data") pod "0b2dd611-2516-4714-8a41-c7ebc3fcc045" (UID: "0b2dd611-2516-4714-8a41-c7ebc3fcc045"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.149953 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156183 4706 generic.go:334] "Generic (PLEG): container finished" podID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerID="82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156214 4706 generic.go:334] "Generic (PLEG): container finished" podID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerID="aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600" exitCode=0 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156287 4706 generic.go:334] "Generic (PLEG): container finished" podID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerID="3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a" exitCode=143 Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156308 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerDied","Data":"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156334 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerDied","Data":"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156347 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerDied","Data":"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156364 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"3e01c94a-8908-4a33-a45a-41f48b44217b","Type":"ContainerDied","Data":"7ef0c86d56b6bd9ef153ebfa4176fc93f19ba9f13a3aa1b1dcdf85dade3ab0e8"} Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.156421 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.157188 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.164883 4706 scope.go:117] "RemoveContainer" containerID="b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.172091 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.183043 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.206594 4706 scope.go:117] "RemoveContainer" containerID="f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.213258 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.221332 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.226848 4706 scope.go:117] "RemoveContainer" containerID="68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.227240 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e\": container with ID starting with 68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e not found: ID does not exist" containerID="68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.227283 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e"} err="failed to get container status \"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e\": rpc error: code = NotFound desc = could not find container \"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e\": container with ID starting with 68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.227314 4706 scope.go:117] "RemoveContainer" containerID="b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.227594 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121\": container with ID starting with b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121 not found: ID does not exist" containerID="b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.227624 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121"} err="failed to get container status \"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121\": rpc error: code = NotFound desc = could not find container \"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121\": container with ID starting with b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.227646 4706 scope.go:117] "RemoveContainer" containerID="f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.227972 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432\": container with ID starting with f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432 not found: ID does not exist" containerID="f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.227999 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432"} err="failed to get container status \"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432\": rpc error: code = NotFound desc = could not find container \"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432\": container with ID starting with f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.228030 4706 scope.go:117] "RemoveContainer" containerID="68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.228253 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e"} err="failed to get container status \"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e\": rpc error: code = NotFound desc = could not find container \"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e\": container with ID starting with 68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.228274 4706 scope.go:117] "RemoveContainer" containerID="b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.228499 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121"} err="failed to get container status \"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121\": rpc error: code = NotFound desc = could not find container \"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121\": container with ID starting with b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.228537 4706 scope.go:117] "RemoveContainer" containerID="f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.228895 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432"} err="failed to get container status \"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432\": rpc error: code = NotFound desc = could not find container \"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432\": container with ID starting with f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.228914 4706 scope.go:117] "RemoveContainer" containerID="68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.229175 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e"} err="failed to get container status \"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e\": rpc error: code = NotFound desc = could not find container \"68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e\": container with ID starting with 68c2eeb9deaff715518b582efd96ce8628b0a4403f661be2af48bd9b4123f81e not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.229263 4706 scope.go:117] "RemoveContainer" containerID="b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.229476 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121"} err="failed to get container status \"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121\": rpc error: code = NotFound desc = could not find container \"b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121\": container with ID starting with b57343ee90cf3ea887ad9ad4e9b844307cabc885709c7d53b19a004b072c8121 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.229501 4706 scope.go:117] "RemoveContainer" containerID="f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.229699 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432"} err="failed to get container status \"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432\": rpc error: code = NotFound desc = could not find container \"f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432\": container with ID starting with f6a21765ff7ec5e9cb887812c3c88f57fef638db9565a35de590f8e52acda432 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.229722 4706 scope.go:117] "RemoveContainer" containerID="746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230673 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-scripts\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230704 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5qgd\" (UniqueName: \"kubernetes.io/projected/ab67a0f0-0b9e-436e-b62b-07b5f34de477-kube-api-access-n5qgd\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230739 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230756 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-lib-modules\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230797 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-logs\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230820 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-run\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230847 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230882 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-iscsi\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230939 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-httpd-run\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230960 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-nvme\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.230994 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-sys\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231021 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-config-data\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231043 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-dev\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231066 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-var-locks-brick\") pod \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\" (UID: \"ab67a0f0-0b9e-436e-b62b-07b5f34de477\") " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231181 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-logs" (OuterVolumeSpecName: "logs") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231280 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231388 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231406 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231423 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231434 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb8d9\" (UniqueName: \"kubernetes.io/projected/d4312c89-9114-487a-807c-3d5510238c4d-kube-api-access-pb8d9\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231443 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4312c89-9114-487a-807c-3d5510238c4d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231452 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231462 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231470 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231478 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231487 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2dd611-2516-4714-8a41-c7ebc3fcc045-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231494 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231501 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231509 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e01c94a-8908-4a33-a45a-41f48b44217b-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231523 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231531 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231539 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231547 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4312c89-9114-487a-807c-3d5510238c4d-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231556 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231564 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.231571 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.232107 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.232232 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.232308 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.232352 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-run" (OuterVolumeSpecName: "run") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.232375 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-dev" (OuterVolumeSpecName: "dev") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.232396 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-sys" (OuterVolumeSpecName: "sys") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.232406 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.233352 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-scripts" (OuterVolumeSpecName: "scripts") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.233462 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.233803 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab67a0f0-0b9e-436e-b62b-07b5f34de477-kube-api-access-n5qgd" (OuterVolumeSpecName: "kube-api-access-n5qgd") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "kube-api-access-n5qgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.245502 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.246361 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.246869 4706 scope.go:117] "RemoveContainer" containerID="250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.248438 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.259106 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-config-data" (OuterVolumeSpecName: "config-data") pod "d4312c89-9114-487a-807c-3d5510238c4d" (UID: "d4312c89-9114-487a-807c-3d5510238c4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.266908 4706 scope.go:117] "RemoveContainer" containerID="9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.281260 4706 scope.go:117] "RemoveContainer" containerID="746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.281673 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982\": container with ID starting with 746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982 not found: ID does not exist" containerID="746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.281698 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982"} err="failed to get container status \"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982\": rpc error: code = NotFound desc = could not find container \"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982\": container with ID starting with 746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.281719 4706 scope.go:117] "RemoveContainer" containerID="250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.282121 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc\": container with ID starting with 250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc not found: ID does not exist" containerID="250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.282158 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc"} err="failed to get container status \"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc\": rpc error: code = NotFound desc = could not find container \"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc\": container with ID starting with 250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.282185 4706 scope.go:117] "RemoveContainer" containerID="9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.282482 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413\": container with ID starting with 9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413 not found: ID does not exist" containerID="9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.282614 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413"} err="failed to get container status \"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413\": rpc error: code = NotFound desc = could not find container \"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413\": container with ID starting with 9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.282719 4706 scope.go:117] "RemoveContainer" containerID="746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.283100 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982"} err="failed to get container status \"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982\": rpc error: code = NotFound desc = could not find container \"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982\": container with ID starting with 746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.283138 4706 scope.go:117] "RemoveContainer" containerID="250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.283570 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc"} err="failed to get container status \"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc\": rpc error: code = NotFound desc = could not find container \"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc\": container with ID starting with 250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.283670 4706 scope.go:117] "RemoveContainer" containerID="9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.284022 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413"} err="failed to get container status \"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413\": rpc error: code = NotFound desc = could not find container \"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413\": container with ID starting with 9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.284045 4706 scope.go:117] "RemoveContainer" containerID="746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.284402 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982"} err="failed to get container status \"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982\": rpc error: code = NotFound desc = could not find container \"746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982\": container with ID starting with 746962ab60ee7c008abf8e3c5150ddfc84e40d8f5d10808015fe199d3d48b982 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.284425 4706 scope.go:117] "RemoveContainer" containerID="250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.284706 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc"} err="failed to get container status \"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc\": rpc error: code = NotFound desc = could not find container \"250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc\": container with ID starting with 250e16b467ca82cf7fdbb163da1fec4241d4384b9c665661f681d47666a5dfcc not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.284726 4706 scope.go:117] "RemoveContainer" containerID="9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.284983 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413"} err="failed to get container status \"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413\": rpc error: code = NotFound desc = could not find container \"9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413\": container with ID starting with 9c9132b6c8c80cc5b0237754422b96f21c1d393519873b37c657c50666a03413 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.285001 4706 scope.go:117] "RemoveContainer" containerID="59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.295126 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-config-data" (OuterVolumeSpecName: "config-data") pod "ab67a0f0-0b9e-436e-b62b-07b5f34de477" (UID: "ab67a0f0-0b9e-436e-b62b-07b5f34de477"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.301265 4706 scope.go:117] "RemoveContainer" containerID="0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.322484 4706 scope.go:117] "RemoveContainer" containerID="5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332860 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332882 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332892 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332901 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab67a0f0-0b9e-436e-b62b-07b5f34de477-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332909 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5qgd\" (UniqueName: \"kubernetes.io/projected/ab67a0f0-0b9e-436e-b62b-07b5f34de477-kube-api-access-n5qgd\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332938 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332947 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332955 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332967 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332975 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332984 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332992 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4312c89-9114-487a-807c-3d5510238c4d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.332999 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab67a0f0-0b9e-436e-b62b-07b5f34de477-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.333007 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.333018 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ab67a0f0-0b9e-436e-b62b-07b5f34de477-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.342640 4706 scope.go:117] "RemoveContainer" containerID="59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.343095 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751\": container with ID starting with 59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751 not found: ID does not exist" containerID="59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.343134 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751"} err="failed to get container status \"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751\": rpc error: code = NotFound desc = could not find container \"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751\": container with ID starting with 59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.343160 4706 scope.go:117] "RemoveContainer" containerID="0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.343378 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa\": container with ID starting with 0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa not found: ID does not exist" containerID="0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.343404 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa"} err="failed to get container status \"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa\": rpc error: code = NotFound desc = could not find container \"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa\": container with ID starting with 0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.343422 4706 scope.go:117] "RemoveContainer" containerID="5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.343672 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d\": container with ID starting with 5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d not found: ID does not exist" containerID="5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.343779 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d"} err="failed to get container status \"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d\": rpc error: code = NotFound desc = could not find container \"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d\": container with ID starting with 5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.343855 4706 scope.go:117] "RemoveContainer" containerID="59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.348362 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.350510 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.353860 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751"} err="failed to get container status \"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751\": rpc error: code = NotFound desc = could not find container \"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751\": container with ID starting with 59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.353922 4706 scope.go:117] "RemoveContainer" containerID="0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.355153 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa"} err="failed to get container status \"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa\": rpc error: code = NotFound desc = could not find container \"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa\": container with ID starting with 0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.355274 4706 scope.go:117] "RemoveContainer" containerID="5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.356015 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d"} err="failed to get container status \"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d\": rpc error: code = NotFound desc = could not find container \"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d\": container with ID starting with 5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.356119 4706 scope.go:117] "RemoveContainer" containerID="59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.356588 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751"} err="failed to get container status \"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751\": rpc error: code = NotFound desc = could not find container \"59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751\": container with ID starting with 59b69ab60e8f48dcafcc76f96ac174b3af0e5836acd8dd9a4a772f2775c8b751 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.356654 4706 scope.go:117] "RemoveContainer" containerID="0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.357295 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa"} err="failed to get container status \"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa\": rpc error: code = NotFound desc = could not find container \"0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa\": container with ID starting with 0682ee4751553fb9745e5b6a6a39e02009f3dfac8d6105ae42abb15a53330baa not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.357324 4706 scope.go:117] "RemoveContainer" containerID="5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.357696 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d"} err="failed to get container status \"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d\": rpc error: code = NotFound desc = could not find container \"5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d\": container with ID starting with 5bb9f345e3ea6960907213a1b848862b52cc6aab966ee7c30fdcaa91c18f676d not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.357771 4706 scope.go:117] "RemoveContainer" containerID="82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.387426 4706 scope.go:117] "RemoveContainer" containerID="aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.405990 4706 scope.go:117] "RemoveContainer" containerID="3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.425649 4706 scope.go:117] "RemoveContainer" containerID="82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.426038 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13\": container with ID starting with 82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13 not found: ID does not exist" containerID="82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426078 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13"} err="failed to get container status \"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13\": rpc error: code = NotFound desc = could not find container \"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13\": container with ID starting with 82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426098 4706 scope.go:117] "RemoveContainer" containerID="aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.426386 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600\": container with ID starting with aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600 not found: ID does not exist" containerID="aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426405 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600"} err="failed to get container status \"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600\": rpc error: code = NotFound desc = could not find container \"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600\": container with ID starting with aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426418 4706 scope.go:117] "RemoveContainer" containerID="3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a" Nov 27 07:28:06 crc kubenswrapper[4706]: E1127 07:28:06.426630 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a\": container with ID starting with 3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a not found: ID does not exist" containerID="3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426649 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a"} err="failed to get container status \"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a\": rpc error: code = NotFound desc = could not find container \"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a\": container with ID starting with 3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426660 4706 scope.go:117] "RemoveContainer" containerID="82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426876 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13"} err="failed to get container status \"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13\": rpc error: code = NotFound desc = could not find container \"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13\": container with ID starting with 82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.426893 4706 scope.go:117] "RemoveContainer" containerID="aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427075 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600"} err="failed to get container status \"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600\": rpc error: code = NotFound desc = could not find container \"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600\": container with ID starting with aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427096 4706 scope.go:117] "RemoveContainer" containerID="3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427313 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a"} err="failed to get container status \"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a\": rpc error: code = NotFound desc = could not find container \"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a\": container with ID starting with 3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427356 4706 scope.go:117] "RemoveContainer" containerID="82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427597 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13"} err="failed to get container status \"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13\": rpc error: code = NotFound desc = could not find container \"82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13\": container with ID starting with 82effaa9ef701977ceb4af7ba21b47aad3636a8eca8a90e643504d0c9a27cc13 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427617 4706 scope.go:117] "RemoveContainer" containerID="aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427830 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600"} err="failed to get container status \"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600\": rpc error: code = NotFound desc = could not find container \"aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600\": container with ID starting with aac1d4a25119cb5ceac3b5ff9246687c26c4bc0fad327392c46e7b615a31f600 not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.427847 4706 scope.go:117] "RemoveContainer" containerID="3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.428060 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a"} err="failed to get container status \"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a\": rpc error: code = NotFound desc = could not find container \"3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a\": container with ID starting with 3535ac9e46866d7f8b3a1bdb67f51015d259c609b161ad4a43e108d7cc02517a not found: ID does not exist" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.433979 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.433998 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.471211 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.476624 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.482370 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.489668 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.495152 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.500991 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.788889 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" path="/var/lib/kubelet/pods/0b2dd611-2516-4714-8a41-c7ebc3fcc045/volumes" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.789754 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" path="/var/lib/kubelet/pods/3e01c94a-8908-4a33-a45a-41f48b44217b/volumes" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.791062 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" path="/var/lib/kubelet/pods/ab67a0f0-0b9e-436e-b62b-07b5f34de477/volumes" Nov 27 07:28:06 crc kubenswrapper[4706]: I1127 07:28:06.792032 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4312c89-9114-487a-807c-3d5510238c4d" path="/var/lib/kubelet/pods/d4312c89-9114-487a-807c-3d5510238c4d/volumes" Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.196302 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.197066 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-log" containerID="cri-o://69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a" gracePeriod=30 Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.197166 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-api" containerID="cri-o://0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188" gracePeriod=30 Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.197193 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-httpd" containerID="cri-o://6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa" gracePeriod=30 Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.760129 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.761126 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-log" containerID="cri-o://66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b" gracePeriod=30 Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.761429 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-api" containerID="cri-o://efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b" gracePeriod=30 Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.761456 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-httpd" containerID="cri-o://25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9" gracePeriod=30 Nov 27 07:28:07 crc kubenswrapper[4706]: I1127 07:28:07.970844 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064539 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-config-data\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064643 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-scripts\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064684 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-logs\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064719 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-var-locks-brick\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064749 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-iscsi\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064781 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97tdk\" (UniqueName: \"kubernetes.io/projected/3c6cf53c-105b-4caa-9c94-f7363752c64a-kube-api-access-97tdk\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064840 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-run\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064855 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064897 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-nvme\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064918 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-run" (OuterVolumeSpecName: "run") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064926 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064944 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.064964 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065001 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065050 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-dev\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065086 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-lib-modules\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065135 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-dev" (OuterVolumeSpecName: "dev") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065169 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-sys\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065252 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-httpd-run\") pod \"3c6cf53c-105b-4caa-9c94-f7363752c64a\" (UID: \"3c6cf53c-105b-4caa-9c94-f7363752c64a\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065288 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065314 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-sys" (OuterVolumeSpecName: "sys") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065370 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-logs" (OuterVolumeSpecName: "logs") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065713 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065800 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065816 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065828 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065839 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065851 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065862 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065873 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c6cf53c-105b-4caa-9c94-f7363752c64a-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.065884 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.069699 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.070592 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c6cf53c-105b-4caa-9c94-f7363752c64a-kube-api-access-97tdk" (OuterVolumeSpecName: "kube-api-access-97tdk") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "kube-api-access-97tdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.071106 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-scripts" (OuterVolumeSpecName: "scripts") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.071524 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.152211 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-config-data" (OuterVolumeSpecName: "config-data") pod "3c6cf53c-105b-4caa-9c94-f7363752c64a" (UID: "3c6cf53c-105b-4caa-9c94-f7363752c64a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.167737 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.182435 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.180730 4706 generic.go:334] "Generic (PLEG): container finished" podID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerID="0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188" exitCode=0 Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.180843 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.182551 4706 generic.go:334] "Generic (PLEG): container finished" podID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerID="6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa" exitCode=0 Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.182593 4706 generic.go:334] "Generic (PLEG): container finished" podID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerID="69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a" exitCode=143 Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.182466 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c6cf53c-105b-4caa-9c94-f7363752c64a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.180751 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerDied","Data":"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188"} Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.182884 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.182915 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerDied","Data":"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa"} Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.182991 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerDied","Data":"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a"} Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.183017 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3c6cf53c-105b-4caa-9c94-f7363752c64a","Type":"ContainerDied","Data":"ed9ba87fcb42e2e2a2a4a7ef195b329f94879dd10fc232dd103bfdd3284d4f8f"} Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.183054 4706 scope.go:117] "RemoveContainer" containerID="0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.183180 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c6cf53c-105b-4caa-9c94-f7363752c64a-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.183204 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97tdk\" (UniqueName: \"kubernetes.io/projected/3c6cf53c-105b-4caa-9c94-f7363752c64a-kube-api-access-97tdk\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.186913 4706 generic.go:334] "Generic (PLEG): container finished" podID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerID="25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9" exitCode=0 Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.186944 4706 generic.go:334] "Generic (PLEG): container finished" podID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerID="66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b" exitCode=143 Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.186946 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerDied","Data":"25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9"} Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.186984 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerDied","Data":"66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b"} Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.189343 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.201267 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.213246 4706 scope.go:117] "RemoveContainer" containerID="6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.278452 4706 scope.go:117] "RemoveContainer" containerID="69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.282753 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.284327 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.284356 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.289329 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.304842 4706 scope.go:117] "RemoveContainer" containerID="0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188" Nov 27 07:28:08 crc kubenswrapper[4706]: E1127 07:28:08.305396 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188\": container with ID starting with 0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188 not found: ID does not exist" containerID="0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.305465 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188"} err="failed to get container status \"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188\": rpc error: code = NotFound desc = could not find container \"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188\": container with ID starting with 0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188 not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.305509 4706 scope.go:117] "RemoveContainer" containerID="6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa" Nov 27 07:28:08 crc kubenswrapper[4706]: E1127 07:28:08.307021 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa\": container with ID starting with 6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa not found: ID does not exist" containerID="6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.307097 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa"} err="failed to get container status \"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa\": rpc error: code = NotFound desc = could not find container \"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa\": container with ID starting with 6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.307131 4706 scope.go:117] "RemoveContainer" containerID="69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a" Nov 27 07:28:08 crc kubenswrapper[4706]: E1127 07:28:08.307551 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a\": container with ID starting with 69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a not found: ID does not exist" containerID="69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.307589 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a"} err="failed to get container status \"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a\": rpc error: code = NotFound desc = could not find container \"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a\": container with ID starting with 69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.307613 4706 scope.go:117] "RemoveContainer" containerID="0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.308174 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188"} err="failed to get container status \"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188\": rpc error: code = NotFound desc = could not find container \"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188\": container with ID starting with 0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188 not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.308418 4706 scope.go:117] "RemoveContainer" containerID="6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.308904 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa"} err="failed to get container status \"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa\": rpc error: code = NotFound desc = could not find container \"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa\": container with ID starting with 6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.308939 4706 scope.go:117] "RemoveContainer" containerID="69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.309239 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a"} err="failed to get container status \"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a\": rpc error: code = NotFound desc = could not find container \"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a\": container with ID starting with 69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.309339 4706 scope.go:117] "RemoveContainer" containerID="0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.309782 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188"} err="failed to get container status \"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188\": rpc error: code = NotFound desc = could not find container \"0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188\": container with ID starting with 0c6addb1bc092fd90cad303f79bb6e0e010f15faccbcecd21080eaa25d27e188 not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.309812 4706 scope.go:117] "RemoveContainer" containerID="6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.310129 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa"} err="failed to get container status \"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa\": rpc error: code = NotFound desc = could not find container \"6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa\": container with ID starting with 6237a281b45e2845aea38ac454bece2f88ed077c2ad1bf4f9e8f29be94fd64aa not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.310237 4706 scope.go:117] "RemoveContainer" containerID="69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.310603 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a"} err="failed to get container status \"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a\": rpc error: code = NotFound desc = could not find container \"69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a\": container with ID starting with 69e26bd52b239c8082a9990009023ff08cb9a86daa8e2493ed29ead3374c253a not found: ID does not exist" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.609174 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.703185 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-dev\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.703578 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-logs\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.703685 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-run\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.703330 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-dev" (OuterVolumeSpecName: "dev") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.703815 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-run" (OuterVolumeSpecName: "run") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.704079 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg9dv\" (UniqueName: \"kubernetes.io/projected/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-kube-api-access-jg9dv\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.704326 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-logs" (OuterVolumeSpecName: "logs") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.704773 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-config-data\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.704908 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-scripts\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.704954 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.705500 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-lib-modules\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.705598 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-sys\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.705665 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-httpd-run\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.706059 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.706203 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-iscsi\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.706368 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-var-locks-brick\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.706442 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-nvme\") pod \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\" (UID: \"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06\") " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.706937 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.706973 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707026 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-sys" (OuterVolumeSpecName: "sys") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707042 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707151 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707451 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707904 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707924 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707937 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707950 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707960 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707972 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707982 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.707991 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.708001 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.709330 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-scripts" (OuterVolumeSpecName: "scripts") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.709897 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.710426 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-kube-api-access-jg9dv" (OuterVolumeSpecName: "kube-api-access-jg9dv") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "kube-api-access-jg9dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.711147 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.788616 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" path="/var/lib/kubelet/pods/3c6cf53c-105b-4caa-9c94-f7363752c64a/volumes" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.809371 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg9dv\" (UniqueName: \"kubernetes.io/projected/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-kube-api-access-jg9dv\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.809402 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.809434 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.809453 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.821287 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-config-data" (OuterVolumeSpecName: "config-data") pod "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" (UID: "8feb8fe8-4ba7-4367-bbf7-d0cf49145e06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.826330 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.839042 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.911091 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.911134 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:08 crc kubenswrapper[4706]: I1127 07:28:08.911146 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.204901 4706 generic.go:334] "Generic (PLEG): container finished" podID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerID="efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b" exitCode=0 Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.204995 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerDied","Data":"efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b"} Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.206039 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8feb8fe8-4ba7-4367-bbf7-d0cf49145e06","Type":"ContainerDied","Data":"f5fcaa586207bc23d8cee7b0d97337dd74c758c1e0200c0ad19629492cf3e795"} Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.206132 4706 scope.go:117] "RemoveContainer" containerID="efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.205059 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.248518 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.256125 4706 scope.go:117] "RemoveContainer" containerID="25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.260658 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.288403 4706 scope.go:117] "RemoveContainer" containerID="66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.312137 4706 scope.go:117] "RemoveContainer" containerID="efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b" Nov 27 07:28:09 crc kubenswrapper[4706]: E1127 07:28:09.312694 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b\": container with ID starting with efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b not found: ID does not exist" containerID="efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.312737 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b"} err="failed to get container status \"efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b\": rpc error: code = NotFound desc = could not find container \"efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b\": container with ID starting with efbede87b0f03468b55a313e4a041157026b676a795693bb49c0d2521f80048b not found: ID does not exist" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.312767 4706 scope.go:117] "RemoveContainer" containerID="25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9" Nov 27 07:28:09 crc kubenswrapper[4706]: E1127 07:28:09.313533 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9\": container with ID starting with 25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9 not found: ID does not exist" containerID="25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.313568 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9"} err="failed to get container status \"25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9\": rpc error: code = NotFound desc = could not find container \"25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9\": container with ID starting with 25c8ab104ddc4a8f5107171784aed279b408147260a682a3d577cebba77cf6d9 not found: ID does not exist" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.313591 4706 scope.go:117] "RemoveContainer" containerID="66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b" Nov 27 07:28:09 crc kubenswrapper[4706]: E1127 07:28:09.313945 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b\": container with ID starting with 66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b not found: ID does not exist" containerID="66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b" Nov 27 07:28:09 crc kubenswrapper[4706]: I1127 07:28:09.313976 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b"} err="failed to get container status \"66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b\": rpc error: code = NotFound desc = could not find container \"66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b\": container with ID starting with 66626c87ffd40bf05785b483359dd415de6542d7f07f50749a6edb7da48d4d6b not found: ID does not exist" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.612881 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6lghz"] Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.621448 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6lghz"] Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.635799 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance1a26-account-delete-cgvfr"] Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636076 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636091 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636111 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636117 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636131 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636139 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636150 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636156 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636166 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636173 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636186 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636192 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636202 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636207 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636236 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636242 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636257 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636264 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636274 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636282 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636298 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636304 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636316 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636322 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636334 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636339 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636345 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636350 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636363 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636382 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636388 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636395 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636403 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636408 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: E1127 07:28:10.636415 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636421 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636529 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636543 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636551 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636557 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636566 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636573 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636582 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636591 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab67a0f0-0b9e-436e-b62b-07b5f34de477" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636598 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636604 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e01c94a-8908-4a33-a45a-41f48b44217b" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636610 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636616 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636623 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636632 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4312c89-9114-487a-807c-3d5510238c4d" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636639 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636647 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-httpd" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636656 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c6cf53c-105b-4caa-9c94-f7363752c64a" containerName="glance-log" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.636665 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2dd611-2516-4714-8a41-c7ebc3fcc045" containerName="glance-api" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.637138 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.650895 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance1a26-account-delete-cgvfr"] Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.736029 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f81b6397-514d-45fc-8a70-6c76822c9498-operator-scripts\") pod \"glance1a26-account-delete-cgvfr\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.736116 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkfd4\" (UniqueName: \"kubernetes.io/projected/f81b6397-514d-45fc-8a70-6c76822c9498-kube-api-access-fkfd4\") pod \"glance1a26-account-delete-cgvfr\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.785130 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8feb8fe8-4ba7-4367-bbf7-d0cf49145e06" path="/var/lib/kubelet/pods/8feb8fe8-4ba7-4367-bbf7-d0cf49145e06/volumes" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.785759 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608" path="/var/lib/kubelet/pods/b1f84bc5-7a15-46b3-9a75-ba3d2c9f2608/volumes" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.837317 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f81b6397-514d-45fc-8a70-6c76822c9498-operator-scripts\") pod \"glance1a26-account-delete-cgvfr\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.837378 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkfd4\" (UniqueName: \"kubernetes.io/projected/f81b6397-514d-45fc-8a70-6c76822c9498-kube-api-access-fkfd4\") pod \"glance1a26-account-delete-cgvfr\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.838169 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f81b6397-514d-45fc-8a70-6c76822c9498-operator-scripts\") pod \"glance1a26-account-delete-cgvfr\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.862360 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkfd4\" (UniqueName: \"kubernetes.io/projected/f81b6397-514d-45fc-8a70-6c76822c9498-kube-api-access-fkfd4\") pod \"glance1a26-account-delete-cgvfr\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:10 crc kubenswrapper[4706]: I1127 07:28:10.952633 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:11 crc kubenswrapper[4706]: I1127 07:28:11.478413 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance1a26-account-delete-cgvfr"] Nov 27 07:28:12 crc kubenswrapper[4706]: I1127 07:28:12.231775 4706 generic.go:334] "Generic (PLEG): container finished" podID="f81b6397-514d-45fc-8a70-6c76822c9498" containerID="0b877983289f5c31f7ded4cf7191cc057ee865bc7c2d08a08bc6f39b8eb3a394" exitCode=0 Nov 27 07:28:12 crc kubenswrapper[4706]: I1127 07:28:12.231825 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" event={"ID":"f81b6397-514d-45fc-8a70-6c76822c9498","Type":"ContainerDied","Data":"0b877983289f5c31f7ded4cf7191cc057ee865bc7c2d08a08bc6f39b8eb3a394"} Nov 27 07:28:12 crc kubenswrapper[4706]: I1127 07:28:12.233365 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" event={"ID":"f81b6397-514d-45fc-8a70-6c76822c9498","Type":"ContainerStarted","Data":"822c725ce2c5215fac464629fb555baac05b1158da80e183c249ec34a4170340"} Nov 27 07:28:13 crc kubenswrapper[4706]: I1127 07:28:13.590468 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:13 crc kubenswrapper[4706]: I1127 07:28:13.684498 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkfd4\" (UniqueName: \"kubernetes.io/projected/f81b6397-514d-45fc-8a70-6c76822c9498-kube-api-access-fkfd4\") pod \"f81b6397-514d-45fc-8a70-6c76822c9498\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " Nov 27 07:28:13 crc kubenswrapper[4706]: I1127 07:28:13.684601 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f81b6397-514d-45fc-8a70-6c76822c9498-operator-scripts\") pod \"f81b6397-514d-45fc-8a70-6c76822c9498\" (UID: \"f81b6397-514d-45fc-8a70-6c76822c9498\") " Nov 27 07:28:13 crc kubenswrapper[4706]: I1127 07:28:13.685451 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81b6397-514d-45fc-8a70-6c76822c9498-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f81b6397-514d-45fc-8a70-6c76822c9498" (UID: "f81b6397-514d-45fc-8a70-6c76822c9498"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:28:13 crc kubenswrapper[4706]: I1127 07:28:13.685810 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f81b6397-514d-45fc-8a70-6c76822c9498-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:13 crc kubenswrapper[4706]: I1127 07:28:13.697406 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f81b6397-514d-45fc-8a70-6c76822c9498-kube-api-access-fkfd4" (OuterVolumeSpecName: "kube-api-access-fkfd4") pod "f81b6397-514d-45fc-8a70-6c76822c9498" (UID: "f81b6397-514d-45fc-8a70-6c76822c9498"). InnerVolumeSpecName "kube-api-access-fkfd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:13 crc kubenswrapper[4706]: I1127 07:28:13.787648 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkfd4\" (UniqueName: \"kubernetes.io/projected/f81b6397-514d-45fc-8a70-6c76822c9498-kube-api-access-fkfd4\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:14 crc kubenswrapper[4706]: I1127 07:28:14.253289 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" event={"ID":"f81b6397-514d-45fc-8a70-6c76822c9498","Type":"ContainerDied","Data":"822c725ce2c5215fac464629fb555baac05b1158da80e183c249ec34a4170340"} Nov 27 07:28:14 crc kubenswrapper[4706]: I1127 07:28:14.253662 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="822c725ce2c5215fac464629fb555baac05b1158da80e183c249ec34a4170340" Nov 27 07:28:14 crc kubenswrapper[4706]: I1127 07:28:14.253375 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance1a26-account-delete-cgvfr" Nov 27 07:28:15 crc kubenswrapper[4706]: I1127 07:28:15.679809 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-q9d6s"] Nov 27 07:28:15 crc kubenswrapper[4706]: I1127 07:28:15.685904 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-q9d6s"] Nov 27 07:28:15 crc kubenswrapper[4706]: I1127 07:28:15.699281 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-1a26-account-create-update-ltfrh"] Nov 27 07:28:15 crc kubenswrapper[4706]: I1127 07:28:15.707410 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance1a26-account-delete-cgvfr"] Nov 27 07:28:15 crc kubenswrapper[4706]: I1127 07:28:15.723203 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-1a26-account-create-update-ltfrh"] Nov 27 07:28:15 crc kubenswrapper[4706]: I1127 07:28:15.725811 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance1a26-account-delete-cgvfr"] Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.593126 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-pghpz"] Nov 27 07:28:16 crc kubenswrapper[4706]: E1127 07:28:16.593944 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81b6397-514d-45fc-8a70-6c76822c9498" containerName="mariadb-account-delete" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.593971 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81b6397-514d-45fc-8a70-6c76822c9498" containerName="mariadb-account-delete" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.594213 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f81b6397-514d-45fc-8a70-6c76822c9498" containerName="mariadb-account-delete" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.594970 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.601272 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-0dca-account-create-update-fbzwd"] Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.602963 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.604621 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.611369 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0dca-account-create-update-fbzwd"] Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.619873 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-pghpz"] Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.640188 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vncf\" (UniqueName: \"kubernetes.io/projected/2e498031-2c37-4d99-a07d-0856b64551b1-kube-api-access-7vncf\") pod \"glance-db-create-pghpz\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.640284 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-operator-scripts\") pod \"glance-0dca-account-create-update-fbzwd\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.640335 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdrgf\" (UniqueName: \"kubernetes.io/projected/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-kube-api-access-fdrgf\") pod \"glance-0dca-account-create-update-fbzwd\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.640368 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e498031-2c37-4d99-a07d-0856b64551b1-operator-scripts\") pod \"glance-db-create-pghpz\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.742298 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vncf\" (UniqueName: \"kubernetes.io/projected/2e498031-2c37-4d99-a07d-0856b64551b1-kube-api-access-7vncf\") pod \"glance-db-create-pghpz\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.742449 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-operator-scripts\") pod \"glance-0dca-account-create-update-fbzwd\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.743900 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-operator-scripts\") pod \"glance-0dca-account-create-update-fbzwd\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.744007 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdrgf\" (UniqueName: \"kubernetes.io/projected/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-kube-api-access-fdrgf\") pod \"glance-0dca-account-create-update-fbzwd\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.744645 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e498031-2c37-4d99-a07d-0856b64551b1-operator-scripts\") pod \"glance-db-create-pghpz\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.746622 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e498031-2c37-4d99-a07d-0856b64551b1-operator-scripts\") pod \"glance-db-create-pghpz\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.774334 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdrgf\" (UniqueName: \"kubernetes.io/projected/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-kube-api-access-fdrgf\") pod \"glance-0dca-account-create-update-fbzwd\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.776753 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vncf\" (UniqueName: \"kubernetes.io/projected/2e498031-2c37-4d99-a07d-0856b64551b1-kube-api-access-7vncf\") pod \"glance-db-create-pghpz\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.789795 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f68290c-e2cc-44c9-a321-afd6fd524542" path="/var/lib/kubelet/pods/5f68290c-e2cc-44c9-a321-afd6fd524542/volumes" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.790832 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2bdc87e-a63e-4edb-a63b-af61c7175160" path="/var/lib/kubelet/pods/d2bdc87e-a63e-4edb-a63b-af61c7175160/volumes" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.791876 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f81b6397-514d-45fc-8a70-6c76822c9498" path="/var/lib/kubelet/pods/f81b6397-514d-45fc-8a70-6c76822c9498/volumes" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.915186 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:16 crc kubenswrapper[4706]: I1127 07:28:16.939828 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:17 crc kubenswrapper[4706]: I1127 07:28:17.475283 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-pghpz"] Nov 27 07:28:17 crc kubenswrapper[4706]: I1127 07:28:17.485942 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0dca-account-create-update-fbzwd"] Nov 27 07:28:18 crc kubenswrapper[4706]: I1127 07:28:18.287916 4706 generic.go:334] "Generic (PLEG): container finished" podID="2e498031-2c37-4d99-a07d-0856b64551b1" containerID="e83fb3ceda05b3ec275792009786ee6c28efc49055f1a64d88d8b83e14459425" exitCode=0 Nov 27 07:28:18 crc kubenswrapper[4706]: I1127 07:28:18.288073 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-pghpz" event={"ID":"2e498031-2c37-4d99-a07d-0856b64551b1","Type":"ContainerDied","Data":"e83fb3ceda05b3ec275792009786ee6c28efc49055f1a64d88d8b83e14459425"} Nov 27 07:28:18 crc kubenswrapper[4706]: I1127 07:28:18.289495 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-pghpz" event={"ID":"2e498031-2c37-4d99-a07d-0856b64551b1","Type":"ContainerStarted","Data":"efe7cc67ef4202b78261818484534ef6c9a25fcaba3e95499ddcf17444f69396"} Nov 27 07:28:18 crc kubenswrapper[4706]: I1127 07:28:18.292894 4706 generic.go:334] "Generic (PLEG): container finished" podID="e2d483cc-de6a-4e3f-88ec-c7fe05209e78" containerID="cfcf7879f8f1b32335b57fefdc9fefa2104a7bdf7df49cca449cbdced5505fac" exitCode=0 Nov 27 07:28:18 crc kubenswrapper[4706]: I1127 07:28:18.292948 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" event={"ID":"e2d483cc-de6a-4e3f-88ec-c7fe05209e78","Type":"ContainerDied","Data":"cfcf7879f8f1b32335b57fefdc9fefa2104a7bdf7df49cca449cbdced5505fac"} Nov 27 07:28:18 crc kubenswrapper[4706]: I1127 07:28:18.293116 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" event={"ID":"e2d483cc-de6a-4e3f-88ec-c7fe05209e78","Type":"ContainerStarted","Data":"ee7bc0ebb19a05fd461abb9f9734260f47848628748696dbe2aa51ab1f921a24"} Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.700322 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.705177 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.895621 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-operator-scripts\") pod \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.895820 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e498031-2c37-4d99-a07d-0856b64551b1-operator-scripts\") pod \"2e498031-2c37-4d99-a07d-0856b64551b1\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.895939 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vncf\" (UniqueName: \"kubernetes.io/projected/2e498031-2c37-4d99-a07d-0856b64551b1-kube-api-access-7vncf\") pod \"2e498031-2c37-4d99-a07d-0856b64551b1\" (UID: \"2e498031-2c37-4d99-a07d-0856b64551b1\") " Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.896003 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdrgf\" (UniqueName: \"kubernetes.io/projected/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-kube-api-access-fdrgf\") pod \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\" (UID: \"e2d483cc-de6a-4e3f-88ec-c7fe05209e78\") " Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.896880 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e498031-2c37-4d99-a07d-0856b64551b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e498031-2c37-4d99-a07d-0856b64551b1" (UID: "2e498031-2c37-4d99-a07d-0856b64551b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.897362 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e2d483cc-de6a-4e3f-88ec-c7fe05209e78" (UID: "e2d483cc-de6a-4e3f-88ec-c7fe05209e78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.901890 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e498031-2c37-4d99-a07d-0856b64551b1-kube-api-access-7vncf" (OuterVolumeSpecName: "kube-api-access-7vncf") pod "2e498031-2c37-4d99-a07d-0856b64551b1" (UID: "2e498031-2c37-4d99-a07d-0856b64551b1"). InnerVolumeSpecName "kube-api-access-7vncf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.903273 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-kube-api-access-fdrgf" (OuterVolumeSpecName: "kube-api-access-fdrgf") pod "e2d483cc-de6a-4e3f-88ec-c7fe05209e78" (UID: "e2d483cc-de6a-4e3f-88ec-c7fe05209e78"). InnerVolumeSpecName "kube-api-access-fdrgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.998291 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e498031-2c37-4d99-a07d-0856b64551b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.998336 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vncf\" (UniqueName: \"kubernetes.io/projected/2e498031-2c37-4d99-a07d-0856b64551b1-kube-api-access-7vncf\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.998356 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdrgf\" (UniqueName: \"kubernetes.io/projected/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-kube-api-access-fdrgf\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:19 crc kubenswrapper[4706]: I1127 07:28:19.998373 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d483cc-de6a-4e3f-88ec-c7fe05209e78-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:20 crc kubenswrapper[4706]: I1127 07:28:20.315185 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-pghpz" event={"ID":"2e498031-2c37-4d99-a07d-0856b64551b1","Type":"ContainerDied","Data":"efe7cc67ef4202b78261818484534ef6c9a25fcaba3e95499ddcf17444f69396"} Nov 27 07:28:20 crc kubenswrapper[4706]: I1127 07:28:20.315238 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efe7cc67ef4202b78261818484534ef6c9a25fcaba3e95499ddcf17444f69396" Nov 27 07:28:20 crc kubenswrapper[4706]: I1127 07:28:20.315254 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pghpz" Nov 27 07:28:20 crc kubenswrapper[4706]: I1127 07:28:20.317163 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" event={"ID":"e2d483cc-de6a-4e3f-88ec-c7fe05209e78","Type":"ContainerDied","Data":"ee7bc0ebb19a05fd461abb9f9734260f47848628748696dbe2aa51ab1f921a24"} Nov 27 07:28:20 crc kubenswrapper[4706]: I1127 07:28:20.317191 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee7bc0ebb19a05fd461abb9f9734260f47848628748696dbe2aa51ab1f921a24" Nov 27 07:28:20 crc kubenswrapper[4706]: I1127 07:28:20.317309 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0dca-account-create-update-fbzwd" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.730089 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-6rg2s"] Nov 27 07:28:21 crc kubenswrapper[4706]: E1127 07:28:21.730888 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e498031-2c37-4d99-a07d-0856b64551b1" containerName="mariadb-database-create" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.730910 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e498031-2c37-4d99-a07d-0856b64551b1" containerName="mariadb-database-create" Nov 27 07:28:21 crc kubenswrapper[4706]: E1127 07:28:21.730949 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d483cc-de6a-4e3f-88ec-c7fe05209e78" containerName="mariadb-account-create-update" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.730964 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d483cc-de6a-4e3f-88ec-c7fe05209e78" containerName="mariadb-account-create-update" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.731199 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e498031-2c37-4d99-a07d-0856b64551b1" containerName="mariadb-database-create" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.731264 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d483cc-de6a-4e3f-88ec-c7fe05209e78" containerName="mariadb-account-create-update" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.731998 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.735930 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.737376 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-7872h" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.747960 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6rg2s"] Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.930310 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdvbr\" (UniqueName: \"kubernetes.io/projected/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-kube-api-access-jdvbr\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.930389 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-db-sync-config-data\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:21 crc kubenswrapper[4706]: I1127 07:28:21.930664 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-config-data\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.033535 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdvbr\" (UniqueName: \"kubernetes.io/projected/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-kube-api-access-jdvbr\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.033854 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-db-sync-config-data\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.034080 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-config-data\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.041584 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-config-data\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.050843 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-db-sync-config-data\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.054939 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdvbr\" (UniqueName: \"kubernetes.io/projected/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-kube-api-access-jdvbr\") pod \"glance-db-sync-6rg2s\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.063487 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.301215 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6rg2s"] Nov 27 07:28:22 crc kubenswrapper[4706]: W1127 07:28:22.305401 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4cd67ae_b3b1_4ffd_8f05_819fdef7968a.slice/crio-041f59380d96357fb505250ba86ce4cd1f72e68eaa59cae7d39c982e19cbcfe9 WatchSource:0}: Error finding container 041f59380d96357fb505250ba86ce4cd1f72e68eaa59cae7d39c982e19cbcfe9: Status 404 returned error can't find the container with id 041f59380d96357fb505250ba86ce4cd1f72e68eaa59cae7d39c982e19cbcfe9 Nov 27 07:28:22 crc kubenswrapper[4706]: I1127 07:28:22.333591 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6rg2s" event={"ID":"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a","Type":"ContainerStarted","Data":"041f59380d96357fb505250ba86ce4cd1f72e68eaa59cae7d39c982e19cbcfe9"} Nov 27 07:28:23 crc kubenswrapper[4706]: I1127 07:28:23.342912 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6rg2s" event={"ID":"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a","Type":"ContainerStarted","Data":"3d2a2366c863bf89a56cda712ace9b140dddaf7bc03f07ce8f28ff89074b529d"} Nov 27 07:28:23 crc kubenswrapper[4706]: I1127 07:28:23.362954 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-6rg2s" podStartSLOduration=2.362935066 podStartE2EDuration="2.362935066s" podCreationTimestamp="2025-11-27 07:28:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:23.3567007 +0000 UTC m=+1187.246291510" watchObservedRunningTime="2025-11-27 07:28:23.362935066 +0000 UTC m=+1187.252525876" Nov 27 07:28:26 crc kubenswrapper[4706]: I1127 07:28:26.368886 4706 generic.go:334] "Generic (PLEG): container finished" podID="e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" containerID="3d2a2366c863bf89a56cda712ace9b140dddaf7bc03f07ce8f28ff89074b529d" exitCode=0 Nov 27 07:28:26 crc kubenswrapper[4706]: I1127 07:28:26.369009 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6rg2s" event={"ID":"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a","Type":"ContainerDied","Data":"3d2a2366c863bf89a56cda712ace9b140dddaf7bc03f07ce8f28ff89074b529d"} Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.695211 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.816753 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdvbr\" (UniqueName: \"kubernetes.io/projected/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-kube-api-access-jdvbr\") pod \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.816855 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-config-data\") pod \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.816939 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-db-sync-config-data\") pod \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\" (UID: \"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a\") " Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.822371 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-kube-api-access-jdvbr" (OuterVolumeSpecName: "kube-api-access-jdvbr") pod "e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" (UID: "e4cd67ae-b3b1-4ffd-8f05-819fdef7968a"). InnerVolumeSpecName "kube-api-access-jdvbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.822938 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" (UID: "e4cd67ae-b3b1-4ffd-8f05-819fdef7968a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.872455 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-config-data" (OuterVolumeSpecName: "config-data") pod "e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" (UID: "e4cd67ae-b3b1-4ffd-8f05-819fdef7968a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.919097 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdvbr\" (UniqueName: \"kubernetes.io/projected/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-kube-api-access-jdvbr\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.919146 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:27 crc kubenswrapper[4706]: I1127 07:28:27.919166 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:28 crc kubenswrapper[4706]: I1127 07:28:28.391871 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6rg2s" event={"ID":"e4cd67ae-b3b1-4ffd-8f05-819fdef7968a","Type":"ContainerDied","Data":"041f59380d96357fb505250ba86ce4cd1f72e68eaa59cae7d39c982e19cbcfe9"} Nov 27 07:28:28 crc kubenswrapper[4706]: I1127 07:28:28.391934 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6rg2s" Nov 27 07:28:28 crc kubenswrapper[4706]: I1127 07:28:28.391949 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="041f59380d96357fb505250ba86ce4cd1f72e68eaa59cae7d39c982e19cbcfe9" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.448942 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:28:29 crc kubenswrapper[4706]: E1127 07:28:29.449522 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" containerName="glance-db-sync" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.449534 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" containerName="glance-db-sync" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.449669 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" containerName="glance-db-sync" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.450305 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.452035 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.452253 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-7872h" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.452328 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.470125 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542446 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-logs\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542549 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542576 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542608 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542645 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542676 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpvk4\" (UniqueName: \"kubernetes.io/projected/413a861f-3b19-4369-84b2-1fe482c99c61-kube-api-access-mpvk4\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542827 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542876 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-dev\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542915 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542938 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-run\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.542957 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-config-data\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.543020 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.543076 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-sys\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.543177 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-scripts\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.644885 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-sys\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.644944 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-scripts\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.644972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-logs\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645013 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645042 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645039 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-sys\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645104 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645178 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645304 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645350 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpvk4\" (UniqueName: \"kubernetes.io/projected/413a861f-3b19-4369-84b2-1fe482c99c61-kube-api-access-mpvk4\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645426 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645449 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645478 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-dev\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645512 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645539 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-run\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645559 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-config-data\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645584 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-dev\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645616 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645621 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-logs\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645639 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-run\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645628 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645657 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645434 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645813 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.645947 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.653357 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-config-data\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.658316 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-scripts\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.663251 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.663790 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpvk4\" (UniqueName: \"kubernetes.io/projected/413a861f-3b19-4369-84b2-1fe482c99c61-kube-api-access-mpvk4\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.667845 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.808970 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.876985 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.879054 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.901640 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.902559 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950191 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950249 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950274 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84kzk\" (UniqueName: \"kubernetes.io/projected/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-kube-api-access-84kzk\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950305 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950496 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950535 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-run\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950565 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950643 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950751 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950811 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-sys\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950837 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950864 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950891 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-dev\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:29 crc kubenswrapper[4706]: I1127 07:28:29.950966 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.052711 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-sys\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.052778 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.052811 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.052854 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-dev\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.052932 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.052959 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053006 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053024 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84kzk\" (UniqueName: \"kubernetes.io/projected/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-kube-api-access-84kzk\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053052 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053089 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053103 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-run\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053125 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053144 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053166 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053617 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.053686 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-sys\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054198 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054291 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-dev\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054343 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054431 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054525 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054531 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-run\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054571 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054635 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.054688 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.058966 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.073325 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.085385 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84kzk\" (UniqueName: \"kubernetes.io/projected/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-kube-api-access-84kzk\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.091444 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.094368 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.228332 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.240400 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.420100 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"413a861f-3b19-4369-84b2-1fe482c99c61","Type":"ContainerStarted","Data":"190d1f9665f04f3be09be3030874331c9f620fffee25ce85631a0aeb3e7fed64"} Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.420450 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"413a861f-3b19-4369-84b2-1fe482c99c61","Type":"ContainerStarted","Data":"153e769d33eafa7e459fc0866352d0cfdedea439b2d7da556b0ea234c431f2d1"} Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.461355 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:30 crc kubenswrapper[4706]: I1127 07:28:30.730204 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.433685 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"413a861f-3b19-4369-84b2-1fe482c99c61","Type":"ContainerStarted","Data":"f56640557b0e8ee4286145304263290212404706d15fe0acca9c3e3cf657db86"} Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.437798 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2","Type":"ContainerStarted","Data":"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14"} Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.437851 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2","Type":"ContainerStarted","Data":"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295"} Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.437876 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2","Type":"ContainerStarted","Data":"19f0f769d63b50baeaa9fd12a09058572862e26bf2fc06196513c4e617071058"} Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.438228 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-httpd" containerID="cri-o://ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14" gracePeriod=30 Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.438211 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-log" containerID="cri-o://2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295" gracePeriod=30 Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.473030 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.473003591 podStartE2EDuration="2.473003591s" podCreationTimestamp="2025-11-27 07:28:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:31.465782333 +0000 UTC m=+1195.355373183" watchObservedRunningTime="2025-11-27 07:28:31.473003591 +0000 UTC m=+1195.362594431" Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.513491 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.513464006 podStartE2EDuration="3.513464006s" podCreationTimestamp="2025-11-27 07:28:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:31.505364918 +0000 UTC m=+1195.394955738" watchObservedRunningTime="2025-11-27 07:28:31.513464006 +0000 UTC m=+1195.403054856" Nov 27 07:28:31 crc kubenswrapper[4706]: I1127 07:28:31.921766 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086638 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84kzk\" (UniqueName: \"kubernetes.io/projected/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-kube-api-access-84kzk\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086693 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086715 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-scripts\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086735 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-iscsi\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086797 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086856 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086908 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-lib-modules\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.086926 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-sys\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087273 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087333 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-nvme\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087342 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-sys" (OuterVolumeSpecName: "sys") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087362 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-dev\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087371 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087419 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-config-data\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087470 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-dev" (OuterVolumeSpecName: "dev") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087517 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-run\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.087579 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-run" (OuterVolumeSpecName: "run") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088175 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-logs\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088237 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-var-locks-brick\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088279 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-httpd-run\") pod \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\" (UID: \"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2\") " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088336 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088471 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088488 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088498 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088507 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088515 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088523 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088530 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088594 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.088615 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-logs" (OuterVolumeSpecName: "logs") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.091359 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-kube-api-access-84kzk" (OuterVolumeSpecName: "kube-api-access-84kzk") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "kube-api-access-84kzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.091705 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.091734 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-scripts" (OuterVolumeSpecName: "scripts") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.092759 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.144320 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-config-data" (OuterVolumeSpecName: "config-data") pod "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" (UID: "8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.189793 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.189845 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84kzk\" (UniqueName: \"kubernetes.io/projected/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-kube-api-access-84kzk\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.189897 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.189916 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.189949 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.189969 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.189986 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.202729 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.209058 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.291429 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.291472 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.453735 4706 generic.go:334] "Generic (PLEG): container finished" podID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerID="ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14" exitCode=143 Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.453789 4706 generic.go:334] "Generic (PLEG): container finished" podID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerID="2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295" exitCode=143 Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.453846 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.453913 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2","Type":"ContainerDied","Data":"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14"} Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.454022 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2","Type":"ContainerDied","Data":"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295"} Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.454097 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2","Type":"ContainerDied","Data":"19f0f769d63b50baeaa9fd12a09058572862e26bf2fc06196513c4e617071058"} Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.454488 4706 scope.go:117] "RemoveContainer" containerID="ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.488467 4706 scope.go:117] "RemoveContainer" containerID="2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.520451 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.539271 4706 scope.go:117] "RemoveContainer" containerID="ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14" Nov 27 07:28:32 crc kubenswrapper[4706]: E1127 07:28:32.540103 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14\": container with ID starting with ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14 not found: ID does not exist" containerID="ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.540166 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14"} err="failed to get container status \"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14\": rpc error: code = NotFound desc = could not find container \"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14\": container with ID starting with ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14 not found: ID does not exist" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.540207 4706 scope.go:117] "RemoveContainer" containerID="2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295" Nov 27 07:28:32 crc kubenswrapper[4706]: E1127 07:28:32.540904 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295\": container with ID starting with 2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295 not found: ID does not exist" containerID="2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.540955 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295"} err="failed to get container status \"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295\": rpc error: code = NotFound desc = could not find container \"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295\": container with ID starting with 2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295 not found: ID does not exist" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.540987 4706 scope.go:117] "RemoveContainer" containerID="ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.540910 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.541830 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14"} err="failed to get container status \"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14\": rpc error: code = NotFound desc = could not find container \"ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14\": container with ID starting with ff277d60bee55c2733a4b557ef064874f882ee85edeb3c240dfaf0488ef98b14 not found: ID does not exist" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.541890 4706 scope.go:117] "RemoveContainer" containerID="2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.542605 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295"} err="failed to get container status \"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295\": rpc error: code = NotFound desc = could not find container \"2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295\": container with ID starting with 2456b88ec44d3146d53688e70ba2ac3c44d0fce5451ad322c1e377e39969a295 not found: ID does not exist" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.575904 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:32 crc kubenswrapper[4706]: E1127 07:28:32.576734 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-httpd" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.576762 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-httpd" Nov 27 07:28:32 crc kubenswrapper[4706]: E1127 07:28:32.576793 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-log" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.576804 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-log" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.582519 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-log" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.582571 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" containerName="glance-httpd" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.584684 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.588455 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.588828 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698044 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698119 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-run\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698152 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698192 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698215 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698316 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698346 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698369 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-dev\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698420 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698443 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698536 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698606 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-sys\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698638 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsh24\" (UniqueName: \"kubernetes.io/projected/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-kube-api-access-wsh24\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.698701 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.788461 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2" path="/var/lib/kubelet/pods/8b3332c8-5ecd-4d22-8cb3-26f386c7ccf2/volumes" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.801257 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsh24\" (UniqueName: \"kubernetes.io/projected/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-kube-api-access-wsh24\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.801324 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.801359 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.801675 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.801939 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-run\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802008 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802042 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802060 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802084 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802111 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802129 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-dev\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802147 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802171 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802196 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802236 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-sys\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-sys\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802326 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-run\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802347 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-dev\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802374 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802595 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802607 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802701 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802830 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.802940 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.803400 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.809432 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.810039 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.827520 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsh24\" (UniqueName: \"kubernetes.io/projected/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-kube-api-access-wsh24\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.829728 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.838336 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:32 crc kubenswrapper[4706]: I1127 07:28:32.910699 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:33 crc kubenswrapper[4706]: W1127 07:28:33.389364 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddd69bae_4b28_4839_a5f6_cc2477c3d0e5.slice/crio-3e3c8eda6b34d91f3ff841e49bb6ed90c049b79c216ffb4a6f2970632a3a74f5 WatchSource:0}: Error finding container 3e3c8eda6b34d91f3ff841e49bb6ed90c049b79c216ffb4a6f2970632a3a74f5: Status 404 returned error can't find the container with id 3e3c8eda6b34d91f3ff841e49bb6ed90c049b79c216ffb4a6f2970632a3a74f5 Nov 27 07:28:33 crc kubenswrapper[4706]: I1127 07:28:33.394458 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:28:33 crc kubenswrapper[4706]: I1127 07:28:33.463904 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5","Type":"ContainerStarted","Data":"3e3c8eda6b34d91f3ff841e49bb6ed90c049b79c216ffb4a6f2970632a3a74f5"} Nov 27 07:28:34 crc kubenswrapper[4706]: I1127 07:28:34.477839 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5","Type":"ContainerStarted","Data":"e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb"} Nov 27 07:28:34 crc kubenswrapper[4706]: I1127 07:28:34.477922 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5","Type":"ContainerStarted","Data":"663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e"} Nov 27 07:28:34 crc kubenswrapper[4706]: I1127 07:28:34.506056 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.506013693 podStartE2EDuration="2.506013693s" podCreationTimestamp="2025-11-27 07:28:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:34.501842895 +0000 UTC m=+1198.391433775" watchObservedRunningTime="2025-11-27 07:28:34.506013693 +0000 UTC m=+1198.395604543" Nov 27 07:28:39 crc kubenswrapper[4706]: I1127 07:28:39.809994 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:39 crc kubenswrapper[4706]: I1127 07:28:39.810941 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:39 crc kubenswrapper[4706]: I1127 07:28:39.852479 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:39 crc kubenswrapper[4706]: I1127 07:28:39.884339 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:40 crc kubenswrapper[4706]: I1127 07:28:40.539371 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:40 crc kubenswrapper[4706]: I1127 07:28:40.539905 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:42 crc kubenswrapper[4706]: I1127 07:28:42.419932 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:42 crc kubenswrapper[4706]: I1127 07:28:42.421906 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:28:42 crc kubenswrapper[4706]: I1127 07:28:42.911553 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:42 crc kubenswrapper[4706]: I1127 07:28:42.911880 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:42 crc kubenswrapper[4706]: I1127 07:28:42.939460 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:42 crc kubenswrapper[4706]: I1127 07:28:42.969589 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:43 crc kubenswrapper[4706]: I1127 07:28:43.572750 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:43 crc kubenswrapper[4706]: I1127 07:28:43.572803 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:45 crc kubenswrapper[4706]: I1127 07:28:45.477075 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:45 crc kubenswrapper[4706]: I1127 07:28:45.487074 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.742972 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.745275 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.754008 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.755730 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.802247 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.802770 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.802813 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.903438 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-dev\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.903482 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.903542 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-sys\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.903582 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.903604 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-run\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.903927 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906330 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j45f4\" (UniqueName: \"kubernetes.io/projected/e3f4640f-a262-4013-bdb3-54c673101658-kube-api-access-j45f4\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906457 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-dev\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906484 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906504 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-logs\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906556 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-sys\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906588 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906628 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906652 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906687 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nflvm\" (UniqueName: \"kubernetes.io/projected/ccbba552-66fa-483a-889a-698d28f8b640-kube-api-access-nflvm\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906732 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906818 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.906860 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.907658 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.907816 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-config-data\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.907840 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.907862 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.907894 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-logs\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.907918 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.907954 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-scripts\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.908011 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-config-data\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.908034 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.908122 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-scripts\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.908162 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-run\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.936139 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.989815 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:28:48 crc kubenswrapper[4706]: I1127 07:28:48.991595 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.011344 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012329 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-run\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012404 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-dev\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012446 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012474 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-sys\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012524 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-run\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012553 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j45f4\" (UniqueName: \"kubernetes.io/projected/e3f4640f-a262-4013-bdb3-54c673101658-kube-api-access-j45f4\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012585 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-dev\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012612 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012641 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-logs\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012683 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-sys\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012718 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012771 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012809 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012859 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nflvm\" (UniqueName: \"kubernetes.io/projected/ccbba552-66fa-483a-889a-698d28f8b640-kube-api-access-nflvm\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012898 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012928 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012956 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.012991 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013044 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-config-data\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013065 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013085 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013106 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-logs\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013155 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-scripts\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013182 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-config-data\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013185 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-run\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013205 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013258 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-dev\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013262 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-scripts\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013537 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-dev\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013574 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013617 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013801 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-sys\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013873 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-run\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013957 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.013974 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-logs\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.014005 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-sys\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.014069 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.014113 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.014713 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015061 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015063 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015165 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015252 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015429 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015703 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015997 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015801 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.015730 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-logs\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.022196 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-scripts\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.030986 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.031937 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-scripts\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.032454 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-config-data\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.037501 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-config-data\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.043180 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nflvm\" (UniqueName: \"kubernetes.io/projected/ccbba552-66fa-483a-889a-698d28f8b640-kube-api-access-nflvm\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.046023 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j45f4\" (UniqueName: \"kubernetes.io/projected/e3f4640f-a262-4013-bdb3-54c673101658-kube-api-access-j45f4\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.046128 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.080058 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.080061 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.094337 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.107972 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114346 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114379 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-scripts\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114396 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-dev\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114415 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vmx4\" (UniqueName: \"kubernetes.io/projected/07e1e2f8-f8ef-4071-9bd0-09d369738457-kube-api-access-7vmx4\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114437 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114528 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-dev\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114609 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b8bk\" (UniqueName: \"kubernetes.io/projected/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-kube-api-access-7b8bk\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114632 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-run\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114692 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114725 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-logs\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114750 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-run\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114795 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.114973 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115109 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-scripts\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115276 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-sys\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115362 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115459 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-config-data\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115550 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-config-data\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115672 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115811 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-sys\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.115900 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.116075 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.116159 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.116193 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.116290 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.116330 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-logs\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.116374 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.116482 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.123006 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217576 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217829 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-sys\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217856 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217889 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217914 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217911 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217933 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217954 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-sys\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217980 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.217979 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218029 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218080 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-logs\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218127 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218150 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218192 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218198 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218266 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218148 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218488 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219212 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.218296 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219640 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-logs\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219722 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-scripts\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219751 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-dev\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219790 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vmx4\" (UniqueName: \"kubernetes.io/projected/07e1e2f8-f8ef-4071-9bd0-09d369738457-kube-api-access-7vmx4\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219830 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219875 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-dev\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.219968 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-run\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220002 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b8bk\" (UniqueName: \"kubernetes.io/projected/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-kube-api-access-7b8bk\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220045 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220092 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-logs\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220132 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-run\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220164 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220197 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220281 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-scripts\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220328 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-sys\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220360 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220395 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-config-data\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.220431 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-config-data\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.221124 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.221844 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-sys\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.221994 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-run\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.222054 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.222372 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-logs\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.222426 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.222536 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.223391 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-dev\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.223454 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-run\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.223510 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-dev\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.225602 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.226925 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-config-data\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.231373 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-config-data\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.234463 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-scripts\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.235794 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-scripts\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.245063 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b8bk\" (UniqueName: \"kubernetes.io/projected/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-kube-api-access-7b8bk\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.245239 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vmx4\" (UniqueName: \"kubernetes.io/projected/07e1e2f8-f8ef-4071-9bd0-09d369738457-kube-api-access-7vmx4\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.251770 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.262034 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.262342 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.279546 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.402546 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.409939 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.609645 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.618534 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:28:49 crc kubenswrapper[4706]: W1127 07:28:49.620569 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3f4640f_a262_4013_bdb3_54c673101658.slice/crio-137fa0fbcc1307204affa84114f652e61111082176d48c05039e80090f29112a WatchSource:0}: Error finding container 137fa0fbcc1307204affa84114f652e61111082176d48c05039e80090f29112a: Status 404 returned error can't find the container with id 137fa0fbcc1307204affa84114f652e61111082176d48c05039e80090f29112a Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.700308 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:28:49 crc kubenswrapper[4706]: W1127 07:28:49.703979 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaf1e851_effa_4aa6_ae9d_cbeee98e4daf.slice/crio-15508841bb0a75002b992415652551670d450c83e0e16600b7547755de0d08b4 WatchSource:0}: Error finding container 15508841bb0a75002b992415652551670d450c83e0e16600b7547755de0d08b4: Status 404 returned error can't find the container with id 15508841bb0a75002b992415652551670d450c83e0e16600b7547755de0d08b4 Nov 27 07:28:49 crc kubenswrapper[4706]: I1127 07:28:49.982816 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:28:49 crc kubenswrapper[4706]: W1127 07:28:49.984793 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07e1e2f8_f8ef_4071_9bd0_09d369738457.slice/crio-1ec247195c97731da21fc1c66995e6ebb013dd4c63ed68f747888afd2d877b31 WatchSource:0}: Error finding container 1ec247195c97731da21fc1c66995e6ebb013dd4c63ed68f747888afd2d877b31: Status 404 returned error can't find the container with id 1ec247195c97731da21fc1c66995e6ebb013dd4c63ed68f747888afd2d877b31 Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.642595 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"caf1e851-effa-4aa6-ae9d-cbeee98e4daf","Type":"ContainerStarted","Data":"a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.643239 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"caf1e851-effa-4aa6-ae9d-cbeee98e4daf","Type":"ContainerStarted","Data":"d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.643253 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"caf1e851-effa-4aa6-ae9d-cbeee98e4daf","Type":"ContainerStarted","Data":"15508841bb0a75002b992415652551670d450c83e0e16600b7547755de0d08b4"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.646424 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ccbba552-66fa-483a-889a-698d28f8b640","Type":"ContainerStarted","Data":"a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.646468 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ccbba552-66fa-483a-889a-698d28f8b640","Type":"ContainerStarted","Data":"885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.646481 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ccbba552-66fa-483a-889a-698d28f8b640","Type":"ContainerStarted","Data":"1b95919e743c1a7a44632891edd0f2575eb66a40087dd54d5f0887f573bd550b"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.649530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"07e1e2f8-f8ef-4071-9bd0-09d369738457","Type":"ContainerStarted","Data":"b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.649813 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"07e1e2f8-f8ef-4071-9bd0-09d369738457","Type":"ContainerStarted","Data":"4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.649825 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"07e1e2f8-f8ef-4071-9bd0-09d369738457","Type":"ContainerStarted","Data":"1ec247195c97731da21fc1c66995e6ebb013dd4c63ed68f747888afd2d877b31"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.651251 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e3f4640f-a262-4013-bdb3-54c673101658","Type":"ContainerStarted","Data":"55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.651408 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e3f4640f-a262-4013-bdb3-54c673101658","Type":"ContainerStarted","Data":"154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.651495 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e3f4640f-a262-4013-bdb3-54c673101658","Type":"ContainerStarted","Data":"137fa0fbcc1307204affa84114f652e61111082176d48c05039e80090f29112a"} Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.698570 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=3.69853859 podStartE2EDuration="3.69853859s" podCreationTimestamp="2025-11-27 07:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:50.696922436 +0000 UTC m=+1214.586513246" watchObservedRunningTime="2025-11-27 07:28:50.69853859 +0000 UTC m=+1214.588129400" Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.699795 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.699787891 podStartE2EDuration="3.699787891s" podCreationTimestamp="2025-11-27 07:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:50.67065849 +0000 UTC m=+1214.560249320" watchObservedRunningTime="2025-11-27 07:28:50.699787891 +0000 UTC m=+1214.589378701" Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.738953 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.738933672 podStartE2EDuration="3.738933672s" podCreationTimestamp="2025-11-27 07:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:50.730456033 +0000 UTC m=+1214.620046853" watchObservedRunningTime="2025-11-27 07:28:50.738933672 +0000 UTC m=+1214.628524502" Nov 27 07:28:50 crc kubenswrapper[4706]: I1127 07:28:50.764335 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.76431927 podStartE2EDuration="3.76431927s" podCreationTimestamp="2025-11-27 07:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:28:50.762910134 +0000 UTC m=+1214.652500954" watchObservedRunningTime="2025-11-27 07:28:50.76431927 +0000 UTC m=+1214.653910080" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.109270 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.109869 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.123945 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.124120 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.154662 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.170989 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.171408 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.180127 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.402821 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.403265 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.410385 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.410429 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.427549 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.442234 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.444812 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.446181 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.767887 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.767935 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.767954 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.767973 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.767990 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.768007 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.768027 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:28:59 crc kubenswrapper[4706]: I1127 07:28:59.768348 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.642635 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.644666 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.645516 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.649444 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.688968 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.690767 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.783811 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.783833 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.785188 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:29:01 crc kubenswrapper[4706]: I1127 07:29:01.816508 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:29:02 crc kubenswrapper[4706]: I1127 07:29:02.421463 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:29:02 crc kubenswrapper[4706]: I1127 07:29:02.431099 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:29:02 crc kubenswrapper[4706]: I1127 07:29:02.598422 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:29:02 crc kubenswrapper[4706]: I1127 07:29:02.607113 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.793944 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-log" containerID="cri-o://885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.794075 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-log" containerID="cri-o://154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.794053 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-httpd" containerID="cri-o://a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.794215 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-log" containerID="cri-o://d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.794290 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-httpd" containerID="cri-o://55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.794334 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-httpd" containerID="cri-o://a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.794450 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-log" containerID="cri-o://4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: I1127 07:29:03.794558 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-httpd" containerID="cri-o://b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda" gracePeriod=30 Nov 27 07:29:03 crc kubenswrapper[4706]: E1127 07:29:03.877427 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3f4640f_a262_4013_bdb3_54c673101658.slice/crio-154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3f4640f_a262_4013_bdb3_54c673101658.slice/crio-conmon-154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3.scope\": RecentStats: unable to find data in memory cache]" Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.804706 4706 generic.go:334] "Generic (PLEG): container finished" podID="e3f4640f-a262-4013-bdb3-54c673101658" containerID="154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3" exitCode=143 Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.804781 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e3f4640f-a262-4013-bdb3-54c673101658","Type":"ContainerDied","Data":"154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3"} Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.807950 4706 generic.go:334] "Generic (PLEG): container finished" podID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerID="d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf" exitCode=143 Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.807988 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"caf1e851-effa-4aa6-ae9d-cbeee98e4daf","Type":"ContainerDied","Data":"d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf"} Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.811954 4706 generic.go:334] "Generic (PLEG): container finished" podID="ccbba552-66fa-483a-889a-698d28f8b640" containerID="885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9" exitCode=143 Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.812006 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ccbba552-66fa-483a-889a-698d28f8b640","Type":"ContainerDied","Data":"885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9"} Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.814857 4706 generic.go:334] "Generic (PLEG): container finished" podID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerID="4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154" exitCode=143 Nov 27 07:29:04 crc kubenswrapper[4706]: I1127 07:29:04.814886 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"07e1e2f8-f8ef-4071-9bd0-09d369738457","Type":"ContainerDied","Data":"4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.357939 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.438052 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.438434 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.466959 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-httpd-run\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467004 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-iscsi\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467020 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-run\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467048 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467111 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-logs\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467130 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-nvme\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467148 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-sys\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467162 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-var-locks-brick\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467229 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nflvm\" (UniqueName: \"kubernetes.io/projected/ccbba552-66fa-483a-889a-698d28f8b640-kube-api-access-nflvm\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467251 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-scripts\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467284 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-dev\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467312 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-config-data\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467343 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-lib-modules\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.467364 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ccbba552-66fa-483a-889a-698d28f8b640\" (UID: \"ccbba552-66fa-483a-889a-698d28f8b640\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.469342 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-sys" (OuterVolumeSpecName: "sys") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.471331 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-dev" (OuterVolumeSpecName: "dev") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.471407 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.471486 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.471507 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.473821 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-scripts" (OuterVolumeSpecName: "scripts") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.474057 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-run" (OuterVolumeSpecName: "run") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.474080 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.474313 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.474325 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-logs" (OuterVolumeSpecName: "logs") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.474376 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.477570 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccbba552-66fa-483a-889a-698d28f8b640-kube-api-access-nflvm" (OuterVolumeSpecName: "kube-api-access-nflvm") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "kube-api-access-nflvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.478371 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.508531 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.518701 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-config-data" (OuterVolumeSpecName: "config-data") pod "ccbba552-66fa-483a-889a-698d28f8b640" (UID: "ccbba552-66fa-483a-889a-698d28f8b640"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.568953 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-scripts\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569171 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-nvme\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569349 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j45f4\" (UniqueName: \"kubernetes.io/projected/e3f4640f-a262-4013-bdb3-54c673101658-kube-api-access-j45f4\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569424 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-iscsi\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569496 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-lib-modules\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569566 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-var-locks-brick\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569634 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-logs\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569701 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-lib-modules\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569790 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b8bk\" (UniqueName: \"kubernetes.io/projected/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-kube-api-access-7b8bk\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569629 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569686 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569715 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569733 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569854 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-dev\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569953 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569962 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-logs" (OuterVolumeSpecName: "logs") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.569987 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-httpd-run\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570026 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-logs\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570044 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-dev\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570064 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vmx4\" (UniqueName: \"kubernetes.io/projected/07e1e2f8-f8ef-4071-9bd0-09d369738457-kube-api-access-7vmx4\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570096 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570111 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570143 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-scripts\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570160 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-dev\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570174 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570182 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570202 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-run\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570239 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-logs\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570255 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570286 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-scripts\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570304 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-sys\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570316 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-sys\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570332 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-nvme\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570327 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-logs" (OuterVolumeSpecName: "logs") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570355 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-run\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570375 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-sys\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570397 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-config-data\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570425 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-var-locks-brick\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570452 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-config-data\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570468 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-iscsi\") pod \"07e1e2f8-f8ef-4071-9bd0-09d369738457\" (UID: \"07e1e2f8-f8ef-4071-9bd0-09d369738457\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570482 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-httpd-run\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570501 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-httpd-run\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570514 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570536 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-logs" (OuterVolumeSpecName: "logs") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570548 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-config-data\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570565 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-iscsi\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570585 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-nvme\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570601 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-lib-modules\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570615 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-var-locks-brick\") pod \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\" (UID: \"caf1e851-effa-4aa6-ae9d-cbeee98e4daf\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570629 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-run\") pod \"e3f4640f-a262-4013-bdb3-54c673101658\" (UID: \"e3f4640f-a262-4013-bdb3-54c673101658\") " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.570879 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571059 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-run" (OuterVolumeSpecName: "run") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571073 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-dev" (OuterVolumeSpecName: "dev") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571098 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571114 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571185 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571200 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571212 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nflvm\" (UniqueName: \"kubernetes.io/projected/ccbba552-66fa-483a-889a-698d28f8b640-kube-api-access-nflvm\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571217 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-dev" (OuterVolumeSpecName: "dev") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571240 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571249 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571260 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571268 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbba552-66fa-483a-889a-698d28f8b640-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571270 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-sys" (OuterVolumeSpecName: "sys") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571276 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571285 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571293 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571301 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571309 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571319 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571346 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571359 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571372 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571403 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571415 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571427 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571445 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571455 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571466 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571475 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccbba552-66fa-483a-889a-698d28f8b640-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571494 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571503 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ccbba552-66fa-483a-889a-698d28f8b640-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571512 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07e1e2f8-f8ef-4071-9bd0-09d369738457-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.571612 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.572110 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.572917 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3f4640f-a262-4013-bdb3-54c673101658-kube-api-access-j45f4" (OuterVolumeSpecName: "kube-api-access-j45f4") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "kube-api-access-j45f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.574587 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.574531 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance-cache") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.574735 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-scripts" (OuterVolumeSpecName: "scripts") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575410 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07e1e2f8-f8ef-4071-9bd0-09d369738457-kube-api-access-7vmx4" (OuterVolumeSpecName: "kube-api-access-7vmx4") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "kube-api-access-7vmx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575438 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance-cache") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575453 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-dev" (OuterVolumeSpecName: "dev") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575483 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-sys" (OuterVolumeSpecName: "sys") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575485 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575507 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575508 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575523 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575538 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-run" (OuterVolumeSpecName: "run") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575548 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-sys" (OuterVolumeSpecName: "sys") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575558 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-run" (OuterVolumeSpecName: "run") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575578 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.575998 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance-cache") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.576622 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.577958 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.580555 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-kube-api-access-7b8bk" (OuterVolumeSpecName: "kube-api-access-7b8bk") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "kube-api-access-7b8bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.583051 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-scripts" (OuterVolumeSpecName: "scripts") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.593910 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.597184 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-scripts" (OuterVolumeSpecName: "scripts") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.616550 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.620256 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-config-data" (OuterVolumeSpecName: "config-data") pod "caf1e851-effa-4aa6-ae9d-cbeee98e4daf" (UID: "caf1e851-effa-4aa6-ae9d-cbeee98e4daf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.622726 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-config-data" (OuterVolumeSpecName: "config-data") pod "e3f4640f-a262-4013-bdb3-54c673101658" (UID: "e3f4640f-a262-4013-bdb3-54c673101658"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.633796 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-config-data" (OuterVolumeSpecName: "config-data") pod "07e1e2f8-f8ef-4071-9bd0-09d369738457" (UID: "07e1e2f8-f8ef-4071-9bd0-09d369738457"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673294 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j45f4\" (UniqueName: \"kubernetes.io/projected/e3f4640f-a262-4013-bdb3-54c673101658-kube-api-access-j45f4\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673326 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673337 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b8bk\" (UniqueName: \"kubernetes.io/projected/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-kube-api-access-7b8bk\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673346 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673377 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673387 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673397 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673405 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vmx4\" (UniqueName: \"kubernetes.io/projected/07e1e2f8-f8ef-4071-9bd0-09d369738457-kube-api-access-7vmx4\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673420 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673433 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673442 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673453 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673465 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673473 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673481 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673489 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673498 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673508 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673516 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673524 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673532 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07e1e2f8-f8ef-4071-9bd0-09d369738457-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673540 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07e1e2f8-f8ef-4071-9bd0-09d369738457-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673547 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673555 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f4640f-a262-4013-bdb3-54c673101658-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673568 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673580 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673590 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673601 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673614 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673625 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/caf1e851-effa-4aa6-ae9d-cbeee98e4daf-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673637 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e3f4640f-a262-4013-bdb3-54c673101658-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.673649 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3f4640f-a262-4013-bdb3-54c673101658-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.685516 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.685604 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.685911 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.686530 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.692547 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.700507 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.775372 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.775421 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.775433 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.775443 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.775458 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.775468 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.864630 4706 generic.go:334] "Generic (PLEG): container finished" podID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerID="b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda" exitCode=0 Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.864743 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"07e1e2f8-f8ef-4071-9bd0-09d369738457","Type":"ContainerDied","Data":"b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.864827 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"07e1e2f8-f8ef-4071-9bd0-09d369738457","Type":"ContainerDied","Data":"1ec247195c97731da21fc1c66995e6ebb013dd4c63ed68f747888afd2d877b31"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.864849 4706 scope.go:117] "RemoveContainer" containerID="b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.865004 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.868124 4706 generic.go:334] "Generic (PLEG): container finished" podID="e3f4640f-a262-4013-bdb3-54c673101658" containerID="55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb" exitCode=0 Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.868191 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e3f4640f-a262-4013-bdb3-54c673101658","Type":"ContainerDied","Data":"55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.868238 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e3f4640f-a262-4013-bdb3-54c673101658","Type":"ContainerDied","Data":"137fa0fbcc1307204affa84114f652e61111082176d48c05039e80090f29112a"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.868248 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.871785 4706 generic.go:334] "Generic (PLEG): container finished" podID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerID="a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808" exitCode=0 Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.871832 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"caf1e851-effa-4aa6-ae9d-cbeee98e4daf","Type":"ContainerDied","Data":"a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.872003 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"caf1e851-effa-4aa6-ae9d-cbeee98e4daf","Type":"ContainerDied","Data":"15508841bb0a75002b992415652551670d450c83e0e16600b7547755de0d08b4"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.871871 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.874112 4706 generic.go:334] "Generic (PLEG): container finished" podID="ccbba552-66fa-483a-889a-698d28f8b640" containerID="a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99" exitCode=0 Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.874135 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ccbba552-66fa-483a-889a-698d28f8b640","Type":"ContainerDied","Data":"a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.874152 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"ccbba552-66fa-483a-889a-698d28f8b640","Type":"ContainerDied","Data":"1b95919e743c1a7a44632891edd0f2575eb66a40087dd54d5f0887f573bd550b"} Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.874160 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.886033 4706 scope.go:117] "RemoveContainer" containerID="4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.909760 4706 scope.go:117] "RemoveContainer" containerID="b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda" Nov 27 07:29:07 crc kubenswrapper[4706]: E1127 07:29:07.915367 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda\": container with ID starting with b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda not found: ID does not exist" containerID="b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.915410 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda"} err="failed to get container status \"b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda\": rpc error: code = NotFound desc = could not find container \"b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda\": container with ID starting with b114e0351b36a724d3a039d99d6ec25929965c330466fbbc76c88981f164dfda not found: ID does not exist" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.915436 4706 scope.go:117] "RemoveContainer" containerID="4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154" Nov 27 07:29:07 crc kubenswrapper[4706]: E1127 07:29:07.916084 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154\": container with ID starting with 4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154 not found: ID does not exist" containerID="4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.916134 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154"} err="failed to get container status \"4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154\": rpc error: code = NotFound desc = could not find container \"4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154\": container with ID starting with 4d462381e6c37556fac560f80db1f5a0d85499b1cb04aa276720e6de2eafa154 not found: ID does not exist" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.916163 4706 scope.go:117] "RemoveContainer" containerID="55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.917036 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.931570 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.935632 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.937993 4706 scope.go:117] "RemoveContainer" containerID="154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.946061 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.957365 4706 scope.go:117] "RemoveContainer" containerID="55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb" Nov 27 07:29:07 crc kubenswrapper[4706]: E1127 07:29:07.957992 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb\": container with ID starting with 55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb not found: ID does not exist" containerID="55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.958028 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb"} err="failed to get container status \"55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb\": rpc error: code = NotFound desc = could not find container \"55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb\": container with ID starting with 55019655e3f071ec3a48fef7407f8970e555823a0acc679e40726777f9c93dcb not found: ID does not exist" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.958052 4706 scope.go:117] "RemoveContainer" containerID="154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.958136 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:29:07 crc kubenswrapper[4706]: E1127 07:29:07.958462 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3\": container with ID starting with 154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3 not found: ID does not exist" containerID="154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.958482 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3"} err="failed to get container status \"154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3\": rpc error: code = NotFound desc = could not find container \"154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3\": container with ID starting with 154b09c2c651d66bd0f8592810e1a3ce459a7111dffb7a76e30432600774aff3 not found: ID does not exist" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.958496 4706 scope.go:117] "RemoveContainer" containerID="a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.965887 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.971504 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.977643 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.978809 4706 scope.go:117] "RemoveContainer" containerID="d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.994835 4706 scope.go:117] "RemoveContainer" containerID="a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808" Nov 27 07:29:07 crc kubenswrapper[4706]: E1127 07:29:07.995198 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808\": container with ID starting with a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808 not found: ID does not exist" containerID="a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.995263 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808"} err="failed to get container status \"a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808\": rpc error: code = NotFound desc = could not find container \"a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808\": container with ID starting with a7e7133f2b858ccae9b0b7b00a2641bc64e5b3be57abf570299127756baf7808 not found: ID does not exist" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.995283 4706 scope.go:117] "RemoveContainer" containerID="d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf" Nov 27 07:29:07 crc kubenswrapper[4706]: E1127 07:29:07.995565 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf\": container with ID starting with d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf not found: ID does not exist" containerID="d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.995608 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf"} err="failed to get container status \"d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf\": rpc error: code = NotFound desc = could not find container \"d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf\": container with ID starting with d7b30b82e2dd09843150c6b494519d25eaed6c02052f2ae681639c69785c2cdf not found: ID does not exist" Nov 27 07:29:07 crc kubenswrapper[4706]: I1127 07:29:07.995633 4706 scope.go:117] "RemoveContainer" containerID="a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.018288 4706 scope.go:117] "RemoveContainer" containerID="885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.089864 4706 scope.go:117] "RemoveContainer" containerID="a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99" Nov 27 07:29:08 crc kubenswrapper[4706]: E1127 07:29:08.090315 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99\": container with ID starting with a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99 not found: ID does not exist" containerID="a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.090353 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99"} err="failed to get container status \"a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99\": rpc error: code = NotFound desc = could not find container \"a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99\": container with ID starting with a29479b994def56ef39f903a019bc8872c14359ec464ef473190db0035989a99 not found: ID does not exist" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.090383 4706 scope.go:117] "RemoveContainer" containerID="885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9" Nov 27 07:29:08 crc kubenswrapper[4706]: E1127 07:29:08.090630 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9\": container with ID starting with 885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9 not found: ID does not exist" containerID="885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.090652 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9"} err="failed to get container status \"885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9\": rpc error: code = NotFound desc = could not find container \"885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9\": container with ID starting with 885ab6168c175d43165202837eb9e3030a718409654db8ca5930b216958f80d9 not found: ID does not exist" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.785774 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" path="/var/lib/kubelet/pods/07e1e2f8-f8ef-4071-9bd0-09d369738457/volumes" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.786452 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" path="/var/lib/kubelet/pods/caf1e851-effa-4aa6-ae9d-cbeee98e4daf/volumes" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.787053 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccbba552-66fa-483a-889a-698d28f8b640" path="/var/lib/kubelet/pods/ccbba552-66fa-483a-889a-698d28f8b640/volumes" Nov 27 07:29:08 crc kubenswrapper[4706]: I1127 07:29:08.788056 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3f4640f-a262-4013-bdb3-54c673101658" path="/var/lib/kubelet/pods/e3f4640f-a262-4013-bdb3-54c673101658/volumes" Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.092406 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.093016 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-httpd" containerID="cri-o://e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb" gracePeriod=30 Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.093202 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-log" containerID="cri-o://663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e" gracePeriod=30 Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.638298 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.638612 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-log" containerID="cri-o://190d1f9665f04f3be09be3030874331c9f620fffee25ce85631a0aeb3e7fed64" gracePeriod=30 Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.638769 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-httpd" containerID="cri-o://f56640557b0e8ee4286145304263290212404706d15fe0acca9c3e3cf657db86" gracePeriod=30 Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.897010 4706 generic.go:334] "Generic (PLEG): container finished" podID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerID="663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e" exitCode=143 Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.897045 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5","Type":"ContainerDied","Data":"663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e"} Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.898931 4706 generic.go:334] "Generic (PLEG): container finished" podID="413a861f-3b19-4369-84b2-1fe482c99c61" containerID="190d1f9665f04f3be09be3030874331c9f620fffee25ce85631a0aeb3e7fed64" exitCode=143 Nov 27 07:29:09 crc kubenswrapper[4706]: I1127 07:29:09.898984 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"413a861f-3b19-4369-84b2-1fe482c99c61","Type":"ContainerDied","Data":"190d1f9665f04f3be09be3030874331c9f620fffee25ce85631a0aeb3e7fed64"} Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.664429 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750005 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750086 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-dev\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750121 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750172 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-dev" (OuterVolumeSpecName: "dev") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750205 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsh24\" (UniqueName: \"kubernetes.io/projected/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-kube-api-access-wsh24\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750275 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-run\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750308 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-lib-modules\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750348 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-scripts\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750398 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-iscsi\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750429 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-run" (OuterVolumeSpecName: "run") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750445 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-config-data\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750469 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750529 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-httpd-run\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750554 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-nvme\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750582 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-sys\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750633 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-var-locks-brick\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750657 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750828 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-logs\") pod \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\" (UID: \"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5\") " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750662 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750694 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-sys" (OuterVolumeSpecName: "sys") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750717 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.750863 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751171 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-logs" (OuterVolumeSpecName: "logs") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751585 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751609 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751626 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751641 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751654 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751667 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751680 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751693 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.751705 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.759129 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.759311 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-scripts" (OuterVolumeSpecName: "scripts") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.762623 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.767477 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-kube-api-access-wsh24" (OuterVolumeSpecName: "kube-api-access-wsh24") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "kube-api-access-wsh24". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.809154 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-config-data" (OuterVolumeSpecName: "config-data") pod "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" (UID: "ddd69bae-4b28-4839-a5f6-cc2477c3d0e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.853558 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.853597 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.853610 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsh24\" (UniqueName: \"kubernetes.io/projected/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-kube-api-access-wsh24\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.853620 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.853629 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.876248 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.885491 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.926865 4706 generic.go:334] "Generic (PLEG): container finished" podID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerID="e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb" exitCode=0 Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.926956 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.926946 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5","Type":"ContainerDied","Data":"e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb"} Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.927113 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"ddd69bae-4b28-4839-a5f6-cc2477c3d0e5","Type":"ContainerDied","Data":"3e3c8eda6b34d91f3ff841e49bb6ed90c049b79c216ffb4a6f2970632a3a74f5"} Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.927148 4706 scope.go:117] "RemoveContainer" containerID="e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.936183 4706 generic.go:334] "Generic (PLEG): container finished" podID="413a861f-3b19-4369-84b2-1fe482c99c61" containerID="f56640557b0e8ee4286145304263290212404706d15fe0acca9c3e3cf657db86" exitCode=0 Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.936236 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"413a861f-3b19-4369-84b2-1fe482c99c61","Type":"ContainerDied","Data":"f56640557b0e8ee4286145304263290212404706d15fe0acca9c3e3cf657db86"} Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.955304 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.955336 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.959802 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.965589 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:29:12 crc kubenswrapper[4706]: I1127 07:29:12.966673 4706 scope.go:117] "RemoveContainer" containerID="663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.017740 4706 scope.go:117] "RemoveContainer" containerID="e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb" Nov 27 07:29:13 crc kubenswrapper[4706]: E1127 07:29:13.024684 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb\": container with ID starting with e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb not found: ID does not exist" containerID="e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.024733 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb"} err="failed to get container status \"e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb\": rpc error: code = NotFound desc = could not find container \"e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb\": container with ID starting with e2e25b8a8a690b66fdb83c0a14c448ed9af2fd4dd9d55cb31999887e12465ceb not found: ID does not exist" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.024759 4706 scope.go:117] "RemoveContainer" containerID="663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e" Nov 27 07:29:13 crc kubenswrapper[4706]: E1127 07:29:13.027643 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e\": container with ID starting with 663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e not found: ID does not exist" containerID="663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.027678 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e"} err="failed to get container status \"663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e\": rpc error: code = NotFound desc = could not find container \"663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e\": container with ID starting with 663f8daff9d7d411503496274af1b0d05f48532a1cf25b611bd9f0b3f11e900e not found: ID does not exist" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.100273 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.157864 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-logs\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.157924 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-lib-modules\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.157952 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-sys\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.157970 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-dev\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.157987 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-nvme\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158002 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-iscsi\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158024 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-scripts\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158045 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158066 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-run\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158110 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-var-locks-brick\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158131 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpvk4\" (UniqueName: \"kubernetes.io/projected/413a861f-3b19-4369-84b2-1fe482c99c61-kube-api-access-mpvk4\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158148 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158174 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-httpd-run\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158271 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-config-data\") pod \"413a861f-3b19-4369-84b2-1fe482c99c61\" (UID: \"413a861f-3b19-4369-84b2-1fe482c99c61\") " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158871 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-sys" (OuterVolumeSpecName: "sys") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.158994 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.159034 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.159064 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-dev" (OuterVolumeSpecName: "dev") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.159138 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-logs" (OuterVolumeSpecName: "logs") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.159303 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.159330 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-run" (OuterVolumeSpecName: "run") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.159477 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.159511 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.167553 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-scripts" (OuterVolumeSpecName: "scripts") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.168366 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.168505 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/413a861f-3b19-4369-84b2-1fe482c99c61-kube-api-access-mpvk4" (OuterVolumeSpecName: "kube-api-access-mpvk4") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "kube-api-access-mpvk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.174417 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.205634 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-config-data" (OuterVolumeSpecName: "config-data") pod "413a861f-3b19-4369-84b2-1fe482c99c61" (UID: "413a861f-3b19-4369-84b2-1fe482c99c61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260441 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260475 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpvk4\" (UniqueName: \"kubernetes.io/projected/413a861f-3b19-4369-84b2-1fe482c99c61-kube-api-access-mpvk4\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260510 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260521 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260530 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260540 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/413a861f-3b19-4369-84b2-1fe482c99c61-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260548 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260556 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260564 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260572 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260580 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260589 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/413a861f-3b19-4369-84b2-1fe482c99c61-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260603 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.260612 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/413a861f-3b19-4369-84b2-1fe482c99c61-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.287337 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.287730 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.362144 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.362186 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.944573 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"413a861f-3b19-4369-84b2-1fe482c99c61","Type":"ContainerDied","Data":"153e769d33eafa7e459fc0866352d0cfdedea439b2d7da556b0ea234c431f2d1"} Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.944599 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.945017 4706 scope.go:117] "RemoveContainer" containerID="f56640557b0e8ee4286145304263290212404706d15fe0acca9c3e3cf657db86" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.982065 4706 scope.go:117] "RemoveContainer" containerID="190d1f9665f04f3be09be3030874331c9f620fffee25ce85631a0aeb3e7fed64" Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.982954 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:29:13 crc kubenswrapper[4706]: I1127 07:29:13.987897 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:29:14 crc kubenswrapper[4706]: E1127 07:29:14.065028 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod413a861f_3b19_4369_84b2_1fe482c99c61.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod413a861f_3b19_4369_84b2_1fe482c99c61.slice/crio-153e769d33eafa7e459fc0866352d0cfdedea439b2d7da556b0ea234c431f2d1\": RecentStats: unable to find data in memory cache]" Nov 27 07:29:14 crc kubenswrapper[4706]: I1127 07:29:14.793280 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" path="/var/lib/kubelet/pods/413a861f-3b19-4369-84b2-1fe482c99c61/volumes" Nov 27 07:29:14 crc kubenswrapper[4706]: I1127 07:29:14.794662 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" path="/var/lib/kubelet/pods/ddd69bae-4b28-4839-a5f6-cc2477c3d0e5/volumes" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.178078 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.179409 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.405471 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6rg2s"] Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.410959 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6rg2s"] Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485452 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance0dca-account-delete-zhgr5"] Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485700 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485713 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485726 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485731 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485740 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485746 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485761 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485767 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485775 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485781 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485789 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485796 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485810 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485815 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485829 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485835 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485842 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485847 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485855 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485861 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485870 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485876 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: E1127 07:29:15.485885 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.485891 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486010 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486022 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486032 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486040 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbba552-66fa-483a-889a-698d28f8b640" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486048 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486056 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e1e2f8-f8ef-4071-9bd0-09d369738457" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486064 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486073 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3f4640f-a262-4013-bdb3-54c673101658" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486080 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486090 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf1e851-effa-4aa6-ae9d-cbeee98e4daf" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486098 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="413a861f-3b19-4369-84b2-1fe482c99c61" containerName="glance-httpd" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486107 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd69bae-4b28-4839-a5f6-cc2477c3d0e5" containerName="glance-log" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.486535 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.505427 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance0dca-account-delete-zhgr5"] Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.599451 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf5m7\" (UniqueName: \"kubernetes.io/projected/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-kube-api-access-cf5m7\") pod \"glance0dca-account-delete-zhgr5\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.599509 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-operator-scripts\") pod \"glance0dca-account-delete-zhgr5\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.700391 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf5m7\" (UniqueName: \"kubernetes.io/projected/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-kube-api-access-cf5m7\") pod \"glance0dca-account-delete-zhgr5\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.700461 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-operator-scripts\") pod \"glance0dca-account-delete-zhgr5\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.701412 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-operator-scripts\") pod \"glance0dca-account-delete-zhgr5\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.718459 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf5m7\" (UniqueName: \"kubernetes.io/projected/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-kube-api-access-cf5m7\") pod \"glance0dca-account-delete-zhgr5\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:15 crc kubenswrapper[4706]: I1127 07:29:15.806301 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:16 crc kubenswrapper[4706]: I1127 07:29:16.284824 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance0dca-account-delete-zhgr5"] Nov 27 07:29:16 crc kubenswrapper[4706]: I1127 07:29:16.785414 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4cd67ae-b3b1-4ffd-8f05-819fdef7968a" path="/var/lib/kubelet/pods/e4cd67ae-b3b1-4ffd-8f05-819fdef7968a/volumes" Nov 27 07:29:16 crc kubenswrapper[4706]: I1127 07:29:16.974204 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" event={"ID":"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257","Type":"ContainerStarted","Data":"0a759fba5b0de9a2e4b623f0c831b18f663e33d44f411bb224458f47e7d934db"} Nov 27 07:29:16 crc kubenswrapper[4706]: I1127 07:29:16.974288 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" event={"ID":"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257","Type":"ContainerStarted","Data":"007d9aa1b71674a45b6a4c5f8e9a7cb509eaafc89bbb1d3e4aae4425e4ce6582"} Nov 27 07:29:16 crc kubenswrapper[4706]: I1127 07:29:16.992983 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" podStartSLOduration=1.992967863 podStartE2EDuration="1.992967863s" podCreationTimestamp="2025-11-27 07:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:29:16.988712723 +0000 UTC m=+1240.878303533" watchObservedRunningTime="2025-11-27 07:29:16.992967863 +0000 UTC m=+1240.882558663" Nov 27 07:29:18 crc kubenswrapper[4706]: I1127 07:29:18.995377 4706 generic.go:334] "Generic (PLEG): container finished" podID="0423c4f5-7e8b-4a51-bb60-3f2d53ef9257" containerID="0a759fba5b0de9a2e4b623f0c831b18f663e33d44f411bb224458f47e7d934db" exitCode=0 Nov 27 07:29:18 crc kubenswrapper[4706]: I1127 07:29:18.995449 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" event={"ID":"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257","Type":"ContainerDied","Data":"0a759fba5b0de9a2e4b623f0c831b18f663e33d44f411bb224458f47e7d934db"} Nov 27 07:29:20 crc kubenswrapper[4706]: I1127 07:29:20.296322 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:20 crc kubenswrapper[4706]: I1127 07:29:20.375686 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-operator-scripts\") pod \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " Nov 27 07:29:20 crc kubenswrapper[4706]: I1127 07:29:20.375977 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf5m7\" (UniqueName: \"kubernetes.io/projected/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-kube-api-access-cf5m7\") pod \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\" (UID: \"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257\") " Nov 27 07:29:20 crc kubenswrapper[4706]: I1127 07:29:20.376873 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0423c4f5-7e8b-4a51-bb60-3f2d53ef9257" (UID: "0423c4f5-7e8b-4a51-bb60-3f2d53ef9257"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:29:20 crc kubenswrapper[4706]: I1127 07:29:20.384281 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-kube-api-access-cf5m7" (OuterVolumeSpecName: "kube-api-access-cf5m7") pod "0423c4f5-7e8b-4a51-bb60-3f2d53ef9257" (UID: "0423c4f5-7e8b-4a51-bb60-3f2d53ef9257"). InnerVolumeSpecName "kube-api-access-cf5m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:20 crc kubenswrapper[4706]: I1127 07:29:20.477349 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:20 crc kubenswrapper[4706]: I1127 07:29:20.477457 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf5m7\" (UniqueName: \"kubernetes.io/projected/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257-kube-api-access-cf5m7\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:21 crc kubenswrapper[4706]: I1127 07:29:21.011662 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" event={"ID":"0423c4f5-7e8b-4a51-bb60-3f2d53ef9257","Type":"ContainerDied","Data":"007d9aa1b71674a45b6a4c5f8e9a7cb509eaafc89bbb1d3e4aae4425e4ce6582"} Nov 27 07:29:21 crc kubenswrapper[4706]: I1127 07:29:21.011732 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0dca-account-delete-zhgr5" Nov 27 07:29:21 crc kubenswrapper[4706]: I1127 07:29:21.011739 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="007d9aa1b71674a45b6a4c5f8e9a7cb509eaafc89bbb1d3e4aae4425e4ce6582" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.520316 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-pghpz"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.532790 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-pghpz"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.538404 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance0dca-account-delete-zhgr5"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.543105 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-0dca-account-create-update-fbzwd"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.548365 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-0dca-account-create-update-fbzwd"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.552576 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance0dca-account-delete-zhgr5"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.603063 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-9x2v2"] Nov 27 07:29:25 crc kubenswrapper[4706]: E1127 07:29:25.603399 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0423c4f5-7e8b-4a51-bb60-3f2d53ef9257" containerName="mariadb-account-delete" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.603418 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0423c4f5-7e8b-4a51-bb60-3f2d53ef9257" containerName="mariadb-account-delete" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.603565 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0423c4f5-7e8b-4a51-bb60-3f2d53ef9257" containerName="mariadb-account-delete" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.604133 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.615258 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9x2v2"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.692675 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4fed3d-33e8-4bdc-a096-1157a74ae784-operator-scripts\") pod \"glance-db-create-9x2v2\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.692748 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54vmx\" (UniqueName: \"kubernetes.io/projected/5b4fed3d-33e8-4bdc-a096-1157a74ae784-kube-api-access-54vmx\") pod \"glance-db-create-9x2v2\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.704573 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-31fa-account-create-update-79nsh"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.705584 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.707808 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.720556 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-31fa-account-create-update-79nsh"] Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.794256 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec594dd5-ee4d-4587-85ca-244798209a0e-operator-scripts\") pod \"glance-31fa-account-create-update-79nsh\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.794321 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2szv\" (UniqueName: \"kubernetes.io/projected/ec594dd5-ee4d-4587-85ca-244798209a0e-kube-api-access-z2szv\") pod \"glance-31fa-account-create-update-79nsh\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.794385 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4fed3d-33e8-4bdc-a096-1157a74ae784-operator-scripts\") pod \"glance-db-create-9x2v2\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.794458 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54vmx\" (UniqueName: \"kubernetes.io/projected/5b4fed3d-33e8-4bdc-a096-1157a74ae784-kube-api-access-54vmx\") pod \"glance-db-create-9x2v2\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.795305 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4fed3d-33e8-4bdc-a096-1157a74ae784-operator-scripts\") pod \"glance-db-create-9x2v2\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.813992 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54vmx\" (UniqueName: \"kubernetes.io/projected/5b4fed3d-33e8-4bdc-a096-1157a74ae784-kube-api-access-54vmx\") pod \"glance-db-create-9x2v2\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.895473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec594dd5-ee4d-4587-85ca-244798209a0e-operator-scripts\") pod \"glance-31fa-account-create-update-79nsh\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.895533 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2szv\" (UniqueName: \"kubernetes.io/projected/ec594dd5-ee4d-4587-85ca-244798209a0e-kube-api-access-z2szv\") pod \"glance-31fa-account-create-update-79nsh\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.896700 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec594dd5-ee4d-4587-85ca-244798209a0e-operator-scripts\") pod \"glance-31fa-account-create-update-79nsh\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.912198 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2szv\" (UniqueName: \"kubernetes.io/projected/ec594dd5-ee4d-4587-85ca-244798209a0e-kube-api-access-z2szv\") pod \"glance-31fa-account-create-update-79nsh\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:25 crc kubenswrapper[4706]: I1127 07:29:25.937247 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:26 crc kubenswrapper[4706]: I1127 07:29:26.018802 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:26 crc kubenswrapper[4706]: I1127 07:29:26.360185 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9x2v2"] Nov 27 07:29:26 crc kubenswrapper[4706]: I1127 07:29:26.473912 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-31fa-account-create-update-79nsh"] Nov 27 07:29:26 crc kubenswrapper[4706]: I1127 07:29:26.787959 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0423c4f5-7e8b-4a51-bb60-3f2d53ef9257" path="/var/lib/kubelet/pods/0423c4f5-7e8b-4a51-bb60-3f2d53ef9257/volumes" Nov 27 07:29:26 crc kubenswrapper[4706]: I1127 07:29:26.788666 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e498031-2c37-4d99-a07d-0856b64551b1" path="/var/lib/kubelet/pods/2e498031-2c37-4d99-a07d-0856b64551b1/volumes" Nov 27 07:29:26 crc kubenswrapper[4706]: I1127 07:29:26.789121 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d483cc-de6a-4e3f-88ec-c7fe05209e78" path="/var/lib/kubelet/pods/e2d483cc-de6a-4e3f-88ec-c7fe05209e78/volumes" Nov 27 07:29:27 crc kubenswrapper[4706]: I1127 07:29:27.085489 4706 generic.go:334] "Generic (PLEG): container finished" podID="5b4fed3d-33e8-4bdc-a096-1157a74ae784" containerID="022b03133d72ddf87c7fa052b05bc029e5613859e2a9efcc37d46fb41348ae96" exitCode=0 Nov 27 07:29:27 crc kubenswrapper[4706]: I1127 07:29:27.085555 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9x2v2" event={"ID":"5b4fed3d-33e8-4bdc-a096-1157a74ae784","Type":"ContainerDied","Data":"022b03133d72ddf87c7fa052b05bc029e5613859e2a9efcc37d46fb41348ae96"} Nov 27 07:29:27 crc kubenswrapper[4706]: I1127 07:29:27.085579 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9x2v2" event={"ID":"5b4fed3d-33e8-4bdc-a096-1157a74ae784","Type":"ContainerStarted","Data":"bbe4a54b2e749a6742f3b0dbb5f0729f45ee3b1f02a9aedd49cd5c2bf16d0ffe"} Nov 27 07:29:27 crc kubenswrapper[4706]: I1127 07:29:27.087656 4706 generic.go:334] "Generic (PLEG): container finished" podID="ec594dd5-ee4d-4587-85ca-244798209a0e" containerID="a819058332255933355b0cc51c35255f77ceca9b20ebb06695156854d9aa72b6" exitCode=0 Nov 27 07:29:27 crc kubenswrapper[4706]: I1127 07:29:27.087701 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" event={"ID":"ec594dd5-ee4d-4587-85ca-244798209a0e","Type":"ContainerDied","Data":"a819058332255933355b0cc51c35255f77ceca9b20ebb06695156854d9aa72b6"} Nov 27 07:29:27 crc kubenswrapper[4706]: I1127 07:29:27.087726 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" event={"ID":"ec594dd5-ee4d-4587-85ca-244798209a0e","Type":"ContainerStarted","Data":"31511cbbc109f97b8aa83dc728e4611ec4aa1d001c6d530477707b91054ae740"} Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.459465 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.466479 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.537920 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54vmx\" (UniqueName: \"kubernetes.io/projected/5b4fed3d-33e8-4bdc-a096-1157a74ae784-kube-api-access-54vmx\") pod \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.537981 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec594dd5-ee4d-4587-85ca-244798209a0e-operator-scripts\") pod \"ec594dd5-ee4d-4587-85ca-244798209a0e\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.538046 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4fed3d-33e8-4bdc-a096-1157a74ae784-operator-scripts\") pod \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\" (UID: \"5b4fed3d-33e8-4bdc-a096-1157a74ae784\") " Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.538888 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec594dd5-ee4d-4587-85ca-244798209a0e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec594dd5-ee4d-4587-85ca-244798209a0e" (UID: "ec594dd5-ee4d-4587-85ca-244798209a0e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.538960 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b4fed3d-33e8-4bdc-a096-1157a74ae784-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5b4fed3d-33e8-4bdc-a096-1157a74ae784" (UID: "5b4fed3d-33e8-4bdc-a096-1157a74ae784"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.539029 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2szv\" (UniqueName: \"kubernetes.io/projected/ec594dd5-ee4d-4587-85ca-244798209a0e-kube-api-access-z2szv\") pod \"ec594dd5-ee4d-4587-85ca-244798209a0e\" (UID: \"ec594dd5-ee4d-4587-85ca-244798209a0e\") " Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.539839 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4fed3d-33e8-4bdc-a096-1157a74ae784-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.539862 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec594dd5-ee4d-4587-85ca-244798209a0e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.544028 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b4fed3d-33e8-4bdc-a096-1157a74ae784-kube-api-access-54vmx" (OuterVolumeSpecName: "kube-api-access-54vmx") pod "5b4fed3d-33e8-4bdc-a096-1157a74ae784" (UID: "5b4fed3d-33e8-4bdc-a096-1157a74ae784"). InnerVolumeSpecName "kube-api-access-54vmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.544452 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec594dd5-ee4d-4587-85ca-244798209a0e-kube-api-access-z2szv" (OuterVolumeSpecName: "kube-api-access-z2szv") pod "ec594dd5-ee4d-4587-85ca-244798209a0e" (UID: "ec594dd5-ee4d-4587-85ca-244798209a0e"). InnerVolumeSpecName "kube-api-access-z2szv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.641009 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54vmx\" (UniqueName: \"kubernetes.io/projected/5b4fed3d-33e8-4bdc-a096-1157a74ae784-kube-api-access-54vmx\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:28 crc kubenswrapper[4706]: I1127 07:29:28.641436 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2szv\" (UniqueName: \"kubernetes.io/projected/ec594dd5-ee4d-4587-85ca-244798209a0e-kube-api-access-z2szv\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:29 crc kubenswrapper[4706]: I1127 07:29:29.112174 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" event={"ID":"ec594dd5-ee4d-4587-85ca-244798209a0e","Type":"ContainerDied","Data":"31511cbbc109f97b8aa83dc728e4611ec4aa1d001c6d530477707b91054ae740"} Nov 27 07:29:29 crc kubenswrapper[4706]: I1127 07:29:29.112265 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-31fa-account-create-update-79nsh" Nov 27 07:29:29 crc kubenswrapper[4706]: I1127 07:29:29.112292 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31511cbbc109f97b8aa83dc728e4611ec4aa1d001c6d530477707b91054ae740" Nov 27 07:29:29 crc kubenswrapper[4706]: I1127 07:29:29.114664 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9x2v2" event={"ID":"5b4fed3d-33e8-4bdc-a096-1157a74ae784","Type":"ContainerDied","Data":"bbe4a54b2e749a6742f3b0dbb5f0729f45ee3b1f02a9aedd49cd5c2bf16d0ffe"} Nov 27 07:29:29 crc kubenswrapper[4706]: I1127 07:29:29.114711 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbe4a54b2e749a6742f3b0dbb5f0729f45ee3b1f02a9aedd49cd5c2bf16d0ffe" Nov 27 07:29:29 crc kubenswrapper[4706]: I1127 07:29:29.114774 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9x2v2" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.840147 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-d92kp"] Nov 27 07:29:30 crc kubenswrapper[4706]: E1127 07:29:30.840405 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4fed3d-33e8-4bdc-a096-1157a74ae784" containerName="mariadb-database-create" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.840417 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4fed3d-33e8-4bdc-a096-1157a74ae784" containerName="mariadb-database-create" Nov 27 07:29:30 crc kubenswrapper[4706]: E1127 07:29:30.840431 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec594dd5-ee4d-4587-85ca-244798209a0e" containerName="mariadb-account-create-update" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.840437 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec594dd5-ee4d-4587-85ca-244798209a0e" containerName="mariadb-account-create-update" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.840576 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4fed3d-33e8-4bdc-a096-1157a74ae784" containerName="mariadb-database-create" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.840590 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec594dd5-ee4d-4587-85ca-244798209a0e" containerName="mariadb-account-create-update" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.841033 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.846949 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-jqt9s" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.847242 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.868698 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d92kp"] Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.976306 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-db-sync-config-data\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.976606 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnv7j\" (UniqueName: \"kubernetes.io/projected/180c49cf-2510-48db-b203-523b3933d5a0-kube-api-access-tnv7j\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:30 crc kubenswrapper[4706]: I1127 07:29:30.976778 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-config-data\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.077902 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnv7j\" (UniqueName: \"kubernetes.io/projected/180c49cf-2510-48db-b203-523b3933d5a0-kube-api-access-tnv7j\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.078172 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-config-data\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.078261 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-db-sync-config-data\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.084477 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-config-data\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.094651 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-db-sync-config-data\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.098921 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnv7j\" (UniqueName: \"kubernetes.io/projected/180c49cf-2510-48db-b203-523b3933d5a0-kube-api-access-tnv7j\") pod \"glance-db-sync-d92kp\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.162552 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:31 crc kubenswrapper[4706]: I1127 07:29:31.621157 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d92kp"] Nov 27 07:29:32 crc kubenswrapper[4706]: I1127 07:29:32.139810 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d92kp" event={"ID":"180c49cf-2510-48db-b203-523b3933d5a0","Type":"ContainerStarted","Data":"cc159708d5c6c7b8617afd9182ff575a6e5aedbaae77cb94d4b5aa74b9dd27d5"} Nov 27 07:29:33 crc kubenswrapper[4706]: I1127 07:29:33.148479 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d92kp" event={"ID":"180c49cf-2510-48db-b203-523b3933d5a0","Type":"ContainerStarted","Data":"360bf9d54cb5135b9a7ee75e00f1e0790e98428f2a4162427097a7f65f4a09c4"} Nov 27 07:29:33 crc kubenswrapper[4706]: I1127 07:29:33.162776 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-d92kp" podStartSLOduration=3.162752313 podStartE2EDuration="3.162752313s" podCreationTimestamp="2025-11-27 07:29:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:29:33.162099891 +0000 UTC m=+1257.051690711" watchObservedRunningTime="2025-11-27 07:29:33.162752313 +0000 UTC m=+1257.052343143" Nov 27 07:29:35 crc kubenswrapper[4706]: I1127 07:29:35.164916 4706 generic.go:334] "Generic (PLEG): container finished" podID="180c49cf-2510-48db-b203-523b3933d5a0" containerID="360bf9d54cb5135b9a7ee75e00f1e0790e98428f2a4162427097a7f65f4a09c4" exitCode=0 Nov 27 07:29:35 crc kubenswrapper[4706]: I1127 07:29:35.164997 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d92kp" event={"ID":"180c49cf-2510-48db-b203-523b3933d5a0","Type":"ContainerDied","Data":"360bf9d54cb5135b9a7ee75e00f1e0790e98428f2a4162427097a7f65f4a09c4"} Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.519119 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.576945 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnv7j\" (UniqueName: \"kubernetes.io/projected/180c49cf-2510-48db-b203-523b3933d5a0-kube-api-access-tnv7j\") pod \"180c49cf-2510-48db-b203-523b3933d5a0\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.577059 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-config-data\") pod \"180c49cf-2510-48db-b203-523b3933d5a0\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.577093 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-db-sync-config-data\") pod \"180c49cf-2510-48db-b203-523b3933d5a0\" (UID: \"180c49cf-2510-48db-b203-523b3933d5a0\") " Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.582895 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180c49cf-2510-48db-b203-523b3933d5a0-kube-api-access-tnv7j" (OuterVolumeSpecName: "kube-api-access-tnv7j") pod "180c49cf-2510-48db-b203-523b3933d5a0" (UID: "180c49cf-2510-48db-b203-523b3933d5a0"). InnerVolumeSpecName "kube-api-access-tnv7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.582954 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "180c49cf-2510-48db-b203-523b3933d5a0" (UID: "180c49cf-2510-48db-b203-523b3933d5a0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.612320 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-config-data" (OuterVolumeSpecName: "config-data") pod "180c49cf-2510-48db-b203-523b3933d5a0" (UID: "180c49cf-2510-48db-b203-523b3933d5a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.678621 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnv7j\" (UniqueName: \"kubernetes.io/projected/180c49cf-2510-48db-b203-523b3933d5a0-kube-api-access-tnv7j\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.678815 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:36 crc kubenswrapper[4706]: I1127 07:29:36.678874 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/180c49cf-2510-48db-b203-523b3933d5a0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:29:37 crc kubenswrapper[4706]: I1127 07:29:37.186042 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-d92kp" event={"ID":"180c49cf-2510-48db-b203-523b3933d5a0","Type":"ContainerDied","Data":"cc159708d5c6c7b8617afd9182ff575a6e5aedbaae77cb94d4b5aa74b9dd27d5"} Nov 27 07:29:37 crc kubenswrapper[4706]: I1127 07:29:37.186086 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc159708d5c6c7b8617afd9182ff575a6e5aedbaae77cb94d4b5aa74b9dd27d5" Nov 27 07:29:37 crc kubenswrapper[4706]: I1127 07:29:37.186120 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-d92kp" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.436028 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:29:38 crc kubenswrapper[4706]: E1127 07:29:38.436962 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180c49cf-2510-48db-b203-523b3933d5a0" containerName="glance-db-sync" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.437001 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="180c49cf-2510-48db-b203-523b3933d5a0" containerName="glance-db-sync" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.437175 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="180c49cf-2510-48db-b203-523b3933d5a0" containerName="glance-db-sync" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.438309 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.441423 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.441520 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-jqt9s" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.448480 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.456635 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506560 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506613 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-config-data\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506635 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-dev\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506659 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-scripts\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506684 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-nvme\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506704 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-lib-modules\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506729 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-run\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506747 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6sxm\" (UniqueName: \"kubernetes.io/projected/a5ff227b-1215-462b-824d-9081fcfb9fff-kube-api-access-m6sxm\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506768 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506789 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-logs\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506825 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506840 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506866 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-sys\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.506884 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-httpd-run\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608302 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608596 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608723 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-sys\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608845 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-httpd-run\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608992 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609105 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-config-data\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609197 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608740 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608889 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609211 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-dev\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609486 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-scripts\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609595 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-nvme\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609684 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-lib-modules\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609766 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-nvme\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609776 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-run\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609886 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6sxm\" (UniqueName: \"kubernetes.io/projected/a5ff227b-1215-462b-824d-9081fcfb9fff-kube-api-access-m6sxm\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609930 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609970 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-logs\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.609435 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-dev\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.608776 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-sys\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.610081 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-lib-modules\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.610163 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.610259 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-run\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.610455 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-logs\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.619615 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-httpd-run\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.626335 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-scripts\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.626739 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-config-data\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.634698 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.637839 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6sxm\" (UniqueName: \"kubernetes.io/projected/a5ff227b-1215-462b-824d-9081fcfb9fff-kube-api-access-m6sxm\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.647857 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:38 crc kubenswrapper[4706]: I1127 07:29:38.793358 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:39 crc kubenswrapper[4706]: I1127 07:29:39.239806 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:29:40 crc kubenswrapper[4706]: I1127 07:29:40.213710 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a5ff227b-1215-462b-824d-9081fcfb9fff","Type":"ContainerStarted","Data":"5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb"} Nov 27 07:29:40 crc kubenswrapper[4706]: I1127 07:29:40.214524 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a5ff227b-1215-462b-824d-9081fcfb9fff","Type":"ContainerStarted","Data":"ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d"} Nov 27 07:29:40 crc kubenswrapper[4706]: I1127 07:29:40.214547 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a5ff227b-1215-462b-824d-9081fcfb9fff","Type":"ContainerStarted","Data":"b3ad6108d735a77dc7f9dc27a201b8d466406b6548ee4b8424b1ba8d6ef8d1db"} Nov 27 07:29:40 crc kubenswrapper[4706]: I1127 07:29:40.249193 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.249161681 podStartE2EDuration="2.249161681s" podCreationTimestamp="2025-11-27 07:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:29:40.243381741 +0000 UTC m=+1264.132972591" watchObservedRunningTime="2025-11-27 07:29:40.249161681 +0000 UTC m=+1264.138752521" Nov 27 07:29:45 crc kubenswrapper[4706]: I1127 07:29:45.178546 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:29:45 crc kubenswrapper[4706]: I1127 07:29:45.179361 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:29:48 crc kubenswrapper[4706]: I1127 07:29:48.794503 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:48 crc kubenswrapper[4706]: I1127 07:29:48.794834 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:48 crc kubenswrapper[4706]: I1127 07:29:48.826309 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:48 crc kubenswrapper[4706]: I1127 07:29:48.852425 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:49 crc kubenswrapper[4706]: I1127 07:29:49.297942 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:49 crc kubenswrapper[4706]: I1127 07:29:49.297994 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:51 crc kubenswrapper[4706]: I1127 07:29:51.400970 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:51 crc kubenswrapper[4706]: I1127 07:29:51.401684 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:29:51 crc kubenswrapper[4706]: I1127 07:29:51.404252 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.365305 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.366682 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.371003 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.372068 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.382865 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.390178 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.478802 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-nvme\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.478933 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kvx7\" (UniqueName: \"kubernetes.io/projected/26d60511-ef2e-4335-af43-816b2d2389fd-kube-api-access-7kvx7\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479025 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-dev\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479049 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479066 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-scripts\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479099 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-httpd-run\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479134 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479165 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-dev\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479191 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479210 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-scripts\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479289 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-logs\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479312 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479328 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-sys\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479356 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-lib-modules\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479373 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-config-data\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479406 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479545 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq4br\" (UniqueName: \"kubernetes.io/projected/8913e171-ae2c-437c-850f-a84fa2918d5b-kube-api-access-nq4br\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479569 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479589 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-run\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479607 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-sys\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479631 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-config-data\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479651 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-logs\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479696 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-nvme\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479732 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-httpd-run\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479758 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479786 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-lib-modules\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479812 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.479826 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-run\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581519 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-httpd-run\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581570 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581598 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-dev\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581624 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581648 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-scripts\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581677 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-logs\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581700 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581728 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-sys\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581745 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581777 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-lib-modules\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581753 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-lib-modules\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-config-data\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581864 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581893 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq4br\" (UniqueName: \"kubernetes.io/projected/8913e171-ae2c-437c-850f-a84fa2918d5b-kube-api-access-nq4br\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581911 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581926 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-run\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581946 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-sys\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581972 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-config-data\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581999 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-logs\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582021 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-nvme\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582026 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-dev\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582049 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-httpd-run\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582071 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-run\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582092 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582124 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-lib-modules\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.581996 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582168 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582170 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-logs\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582189 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-run\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582261 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582277 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-nvme\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582309 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-sys\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582323 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kvx7\" (UniqueName: \"kubernetes.io/projected/26d60511-ef2e-4335-af43-816b2d2389fd-kube-api-access-7kvx7\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582334 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-nvme\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582391 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582412 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-dev\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582412 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-run\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582435 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582430 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-httpd-run\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-scripts\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582506 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582513 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582580 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-dev\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-sys\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582643 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-nvme\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582683 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-lib-modules\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582709 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-httpd-run\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582716 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582839 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.582996 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-logs\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.588509 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-config-data\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.588888 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-config-data\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.589088 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-scripts\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.596135 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-scripts\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.598742 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq4br\" (UniqueName: \"kubernetes.io/projected/8913e171-ae2c-437c-850f-a84fa2918d5b-kube-api-access-nq4br\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.599891 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kvx7\" (UniqueName: \"kubernetes.io/projected/26d60511-ef2e-4335-af43-816b2d2389fd-kube-api-access-7kvx7\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.610301 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.610643 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.613327 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.644960 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.692544 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:29:54 crc kubenswrapper[4706]: I1127 07:29:54.705253 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:29:55 crc kubenswrapper[4706]: I1127 07:29:55.106395 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 27 07:29:55 crc kubenswrapper[4706]: W1127 07:29:55.113158 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8913e171_ae2c_437c_850f_a84fa2918d5b.slice/crio-d868f6d216b800b5f18a8e8900551f2a8f1e6a846926acaabac96c2459e55ab8 WatchSource:0}: Error finding container d868f6d216b800b5f18a8e8900551f2a8f1e6a846926acaabac96c2459e55ab8: Status 404 returned error can't find the container with id d868f6d216b800b5f18a8e8900551f2a8f1e6a846926acaabac96c2459e55ab8 Nov 27 07:29:55 crc kubenswrapper[4706]: I1127 07:29:55.154025 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:29:55 crc kubenswrapper[4706]: W1127 07:29:55.165610 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26d60511_ef2e_4335_af43_816b2d2389fd.slice/crio-f6f2203b0cfbec31580b065a9633cf67c158a40cad0dffb4ace1ac65fd3318ee WatchSource:0}: Error finding container f6f2203b0cfbec31580b065a9633cf67c158a40cad0dffb4ace1ac65fd3318ee: Status 404 returned error can't find the container with id f6f2203b0cfbec31580b065a9633cf67c158a40cad0dffb4ace1ac65fd3318ee Nov 27 07:29:55 crc kubenswrapper[4706]: I1127 07:29:55.342150 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8913e171-ae2c-437c-850f-a84fa2918d5b","Type":"ContainerStarted","Data":"aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0"} Nov 27 07:29:55 crc kubenswrapper[4706]: I1127 07:29:55.342493 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8913e171-ae2c-437c-850f-a84fa2918d5b","Type":"ContainerStarted","Data":"d868f6d216b800b5f18a8e8900551f2a8f1e6a846926acaabac96c2459e55ab8"} Nov 27 07:29:55 crc kubenswrapper[4706]: I1127 07:29:55.344277 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26d60511-ef2e-4335-af43-816b2d2389fd","Type":"ContainerStarted","Data":"b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1"} Nov 27 07:29:55 crc kubenswrapper[4706]: I1127 07:29:55.344322 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26d60511-ef2e-4335-af43-816b2d2389fd","Type":"ContainerStarted","Data":"f6f2203b0cfbec31580b065a9633cf67c158a40cad0dffb4ace1ac65fd3318ee"} Nov 27 07:29:56 crc kubenswrapper[4706]: I1127 07:29:56.352029 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8913e171-ae2c-437c-850f-a84fa2918d5b","Type":"ContainerStarted","Data":"87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98"} Nov 27 07:29:56 crc kubenswrapper[4706]: I1127 07:29:56.364712 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26d60511-ef2e-4335-af43-816b2d2389fd","Type":"ContainerStarted","Data":"f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904"} Nov 27 07:29:56 crc kubenswrapper[4706]: I1127 07:29:56.382555 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-2" podStartSLOduration=3.382535309 podStartE2EDuration="3.382535309s" podCreationTimestamp="2025-11-27 07:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:29:56.372099764 +0000 UTC m=+1280.261690574" watchObservedRunningTime="2025-11-27 07:29:56.382535309 +0000 UTC m=+1280.272126139" Nov 27 07:29:56 crc kubenswrapper[4706]: I1127 07:29:56.407670 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=3.407654128 podStartE2EDuration="3.407654128s" podCreationTimestamp="2025-11-27 07:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:29:56.401433683 +0000 UTC m=+1280.291024483" watchObservedRunningTime="2025-11-27 07:29:56.407654128 +0000 UTC m=+1280.297244938" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.141007 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.142637 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.151875 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.153313 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.154800 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.155293 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.163213 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.164070 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.175304 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.176361 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.180923 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.187019 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.190957 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.224331 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.261733 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4xkm\" (UniqueName: \"kubernetes.io/projected/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-kube-api-access-t4xkm\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.262039 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgb6v\" (UniqueName: \"kubernetes.io/projected/05f94cc7-9f7a-4af5-a9ce-a403adabf163-kube-api-access-hgb6v\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.262144 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.262252 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-secret-volume\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.262370 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/05f94cc7-9f7a-4af5-a9ce-a403adabf163-image-cache-config-data\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.262504 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-config-volume\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.282162 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364511 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-config-volume\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364572 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4xkm\" (UniqueName: \"kubernetes.io/projected/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-kube-api-access-t4xkm\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364635 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-image-cache-config-data\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364661 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgb6v\" (UniqueName: \"kubernetes.io/projected/05f94cc7-9f7a-4af5-a9ce-a403adabf163-kube-api-access-hgb6v\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364708 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zx8x\" (UniqueName: \"kubernetes.io/projected/56de51ce-dc44-4bda-9b13-191c2cddae41-kube-api-access-7zx8x\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364843 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-secret-volume\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364889 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4jcs\" (UniqueName: \"kubernetes.io/projected/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-kube-api-access-b4jcs\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364912 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/05f94cc7-9f7a-4af5-a9ce-a403adabf163-image-cache-config-data\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364932 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364952 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.364969 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/56de51ce-dc44-4bda-9b13-191c2cddae41-image-cache-config-data\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.365996 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-config-volume\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.378799 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-secret-volume\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.379718 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/05f94cc7-9f7a-4af5-a9ce-a403adabf163-image-cache-config-data\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.382052 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgb6v\" (UniqueName: \"kubernetes.io/projected/05f94cc7-9f7a-4af5-a9ce-a403adabf163-kube-api-access-hgb6v\") pod \"glance-cache-glance-default-single-0-cleaner-29403810-xg8rr\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.385245 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.386530 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4xkm\" (UniqueName: \"kubernetes.io/projected/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-kube-api-access-t4xkm\") pod \"collect-profiles-29403810-nddqd\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.391948 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.465942 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-image-cache-config-data\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.466004 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zx8x\" (UniqueName: \"kubernetes.io/projected/56de51ce-dc44-4bda-9b13-191c2cddae41-kube-api-access-7zx8x\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.466051 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4jcs\" (UniqueName: \"kubernetes.io/projected/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-kube-api-access-b4jcs\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.466084 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/56de51ce-dc44-4bda-9b13-191c2cddae41-image-cache-config-data\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.469520 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-image-cache-config-data\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.470623 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/56de51ce-dc44-4bda-9b13-191c2cddae41-image-cache-config-data\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.482839 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zx8x\" (UniqueName: \"kubernetes.io/projected/56de51ce-dc44-4bda-9b13-191c2cddae41-kube-api-access-7zx8x\") pod \"glance-cache-glance-default-single-2-cleaner-29403810-nr9vp\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.483184 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4jcs\" (UniqueName: \"kubernetes.io/projected/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-kube-api-access-b4jcs\") pod \"glance-cache-glance-default-single-1-cleaner-29403810-zgrl6\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.514204 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.526094 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.536647 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.541546 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.774584 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd"] Nov 27 07:30:00 crc kubenswrapper[4706]: I1127 07:30:00.831799 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr"] Nov 27 07:30:00 crc kubenswrapper[4706]: W1127 07:30:00.834628 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05f94cc7_9f7a_4af5_a9ce_a403adabf163.slice/crio-6791fa8179b58da6b8d4ebfdf7e5fdf0d5bd406aabf2d98c8565451b5a706cc1 WatchSource:0}: Error finding container 6791fa8179b58da6b8d4ebfdf7e5fdf0d5bd406aabf2d98c8565451b5a706cc1: Status 404 returned error can't find the container with id 6791fa8179b58da6b8d4ebfdf7e5fdf0d5bd406aabf2d98c8565451b5a706cc1 Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.095481 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6"] Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.145126 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp"] Nov 27 07:30:01 crc kubenswrapper[4706]: W1127 07:30:01.154446 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56de51ce_dc44_4bda_9b13_191c2cddae41.slice/crio-a25c5c95324cbb513658fd6f3c56438bb3c01297c81f62f9569cf0a9075bd385 WatchSource:0}: Error finding container a25c5c95324cbb513658fd6f3c56438bb3c01297c81f62f9569cf0a9075bd385: Status 404 returned error can't find the container with id a25c5c95324cbb513658fd6f3c56438bb3c01297c81f62f9569cf0a9075bd385 Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.407628 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" event={"ID":"05f94cc7-9f7a-4af5-a9ce-a403adabf163","Type":"ContainerStarted","Data":"6791fa8179b58da6b8d4ebfdf7e5fdf0d5bd406aabf2d98c8565451b5a706cc1"} Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.414672 4706 generic.go:334] "Generic (PLEG): container finished" podID="d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8" containerID="5d61da61cc008b02c6eede34f36cb909f1cc61092951e7491f0a64537c455f1e" exitCode=0 Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.414734 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" event={"ID":"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8","Type":"ContainerDied","Data":"5d61da61cc008b02c6eede34f36cb909f1cc61092951e7491f0a64537c455f1e"} Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.414753 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" event={"ID":"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8","Type":"ContainerStarted","Data":"c89e34a18ec1e236648d5c42d3a393c73dd7cbe46c57500a1d9d7fff9e460216"} Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.417299 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" event={"ID":"56de51ce-dc44-4bda-9b13-191c2cddae41","Type":"ContainerStarted","Data":"a25c5c95324cbb513658fd6f3c56438bb3c01297c81f62f9569cf0a9075bd385"} Nov 27 07:30:01 crc kubenswrapper[4706]: I1127 07:30:01.418775 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" event={"ID":"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681","Type":"ContainerStarted","Data":"059ccd2b1036af6c8baf2cf159988e4f82ef434c66c3e0e9481690d3a9e5d0cb"} Nov 27 07:30:02 crc kubenswrapper[4706]: I1127 07:30:02.428676 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" event={"ID":"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681","Type":"ContainerStarted","Data":"f9e493bc129e15e013be1ae53108b37b262a81e065401f360568a1d57fc111a1"} Nov 27 07:30:02 crc kubenswrapper[4706]: I1127 07:30:02.430423 4706 generic.go:334] "Generic (PLEG): container finished" podID="05f94cc7-9f7a-4af5-a9ce-a403adabf163" containerID="da2bef8fec31533ea8254a7be675d2b633117bd57b9252efe993948959dd22e8" exitCode=0 Nov 27 07:30:02 crc kubenswrapper[4706]: I1127 07:30:02.430472 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" event={"ID":"05f94cc7-9f7a-4af5-a9ce-a403adabf163","Type":"ContainerDied","Data":"da2bef8fec31533ea8254a7be675d2b633117bd57b9252efe993948959dd22e8"} Nov 27 07:30:02 crc kubenswrapper[4706]: I1127 07:30:02.432124 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" event={"ID":"56de51ce-dc44-4bda-9b13-191c2cddae41","Type":"ContainerStarted","Data":"e486246d7f6dc59b55b26a87ad194529362f0f9e42095eda2b81f5ee782873ad"} Nov 27 07:30:02 crc kubenswrapper[4706]: I1127 07:30:02.453980 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" podStartSLOduration=2.453941481 podStartE2EDuration="2.453941481s" podCreationTimestamp="2025-11-27 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:02.445806202 +0000 UTC m=+1286.335397012" watchObservedRunningTime="2025-11-27 07:30:02.453941481 +0000 UTC m=+1286.343532281" Nov 27 07:30:02 crc kubenswrapper[4706]: I1127 07:30:02.469988 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" podStartSLOduration=2.469816224 podStartE2EDuration="2.469816224s" podCreationTimestamp="2025-11-27 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:02.468632176 +0000 UTC m=+1286.358222986" watchObservedRunningTime="2025-11-27 07:30:02.469816224 +0000 UTC m=+1286.359407044" Nov 27 07:30:02 crc kubenswrapper[4706]: I1127 07:30:02.856814 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.012798 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4xkm\" (UniqueName: \"kubernetes.io/projected/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-kube-api-access-t4xkm\") pod \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.012889 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-secret-volume\") pod \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.012921 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-config-volume\") pod \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\" (UID: \"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8\") " Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.013595 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-config-volume" (OuterVolumeSpecName: "config-volume") pod "d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8" (UID: "d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.013795 4706 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.024423 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8" (UID: "d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.024509 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-kube-api-access-t4xkm" (OuterVolumeSpecName: "kube-api-access-t4xkm") pod "d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8" (UID: "d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8"). InnerVolumeSpecName "kube-api-access-t4xkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.115331 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4xkm\" (UniqueName: \"kubernetes.io/projected/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-kube-api-access-t4xkm\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.115375 4706 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.439758 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" event={"ID":"d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8","Type":"ContainerDied","Data":"c89e34a18ec1e236648d5c42d3a393c73dd7cbe46c57500a1d9d7fff9e460216"} Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.439806 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c89e34a18ec1e236648d5c42d3a393c73dd7cbe46c57500a1d9d7fff9e460216" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.439877 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403810-nddqd" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.446913 4706 generic.go:334] "Generic (PLEG): container finished" podID="56de51ce-dc44-4bda-9b13-191c2cddae41" containerID="e486246d7f6dc59b55b26a87ad194529362f0f9e42095eda2b81f5ee782873ad" exitCode=0 Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.447003 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" event={"ID":"56de51ce-dc44-4bda-9b13-191c2cddae41","Type":"ContainerDied","Data":"e486246d7f6dc59b55b26a87ad194529362f0f9e42095eda2b81f5ee782873ad"} Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.448923 4706 generic.go:334] "Generic (PLEG): container finished" podID="f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" containerID="f9e493bc129e15e013be1ae53108b37b262a81e065401f360568a1d57fc111a1" exitCode=0 Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.448970 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" event={"ID":"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681","Type":"ContainerDied","Data":"f9e493bc129e15e013be1ae53108b37b262a81e065401f360568a1d57fc111a1"} Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.690601 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.830504 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgb6v\" (UniqueName: \"kubernetes.io/projected/05f94cc7-9f7a-4af5-a9ce-a403adabf163-kube-api-access-hgb6v\") pod \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.830649 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/05f94cc7-9f7a-4af5-a9ce-a403adabf163-image-cache-config-data\") pod \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.830696 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\" (UID: \"05f94cc7-9f7a-4af5-a9ce-a403adabf163\") " Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.836381 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05f94cc7-9f7a-4af5-a9ce-a403adabf163-kube-api-access-hgb6v" (OuterVolumeSpecName: "kube-api-access-hgb6v") pod "05f94cc7-9f7a-4af5-a9ce-a403adabf163" (UID: "05f94cc7-9f7a-4af5-a9ce-a403adabf163"). InnerVolumeSpecName "kube-api-access-hgb6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.836671 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "05f94cc7-9f7a-4af5-a9ce-a403adabf163" (UID: "05f94cc7-9f7a-4af5-a9ce-a403adabf163"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.838458 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05f94cc7-9f7a-4af5-a9ce-a403adabf163-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "05f94cc7-9f7a-4af5-a9ce-a403adabf163" (UID: "05f94cc7-9f7a-4af5-a9ce-a403adabf163"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.932678 4706 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/05f94cc7-9f7a-4af5-a9ce-a403adabf163-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:03 crc kubenswrapper[4706]: I1127 07:30:03.932711 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgb6v\" (UniqueName: \"kubernetes.io/projected/05f94cc7-9f7a-4af5-a9ce-a403adabf163-kube-api-access-hgb6v\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.460780 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" event={"ID":"05f94cc7-9f7a-4af5-a9ce-a403adabf163","Type":"ContainerDied","Data":"6791fa8179b58da6b8d4ebfdf7e5fdf0d5bd406aabf2d98c8565451b5a706cc1"} Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.460873 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.460862 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6791fa8179b58da6b8d4ebfdf7e5fdf0d5bd406aabf2d98c8565451b5a706cc1" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.693907 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.694201 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.706610 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.706658 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.735406 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.753945 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.754384 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.754629 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.793292 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.796131 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.954159 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-image-cache-config-data\") pod \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.954349 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zx8x\" (UniqueName: \"kubernetes.io/projected/56de51ce-dc44-4bda-9b13-191c2cddae41-kube-api-access-7zx8x\") pod \"56de51ce-dc44-4bda-9b13-191c2cddae41\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.954406 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/56de51ce-dc44-4bda-9b13-191c2cddae41-image-cache-config-data\") pod \"56de51ce-dc44-4bda-9b13-191c2cddae41\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.954443 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"56de51ce-dc44-4bda-9b13-191c2cddae41\" (UID: \"56de51ce-dc44-4bda-9b13-191c2cddae41\") " Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.954470 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4jcs\" (UniqueName: \"kubernetes.io/projected/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-kube-api-access-b4jcs\") pod \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.954488 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\" (UID: \"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681\") " Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.959122 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "56de51ce-dc44-4bda-9b13-191c2cddae41" (UID: "56de51ce-dc44-4bda-9b13-191c2cddae41"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.959174 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance-cache") pod "f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" (UID: "f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.972107 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-kube-api-access-b4jcs" (OuterVolumeSpecName: "kube-api-access-b4jcs") pod "f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" (UID: "f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681"). InnerVolumeSpecName "kube-api-access-b4jcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.972108 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56de51ce-dc44-4bda-9b13-191c2cddae41-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "56de51ce-dc44-4bda-9b13-191c2cddae41" (UID: "56de51ce-dc44-4bda-9b13-191c2cddae41"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.972152 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56de51ce-dc44-4bda-9b13-191c2cddae41-kube-api-access-7zx8x" (OuterVolumeSpecName: "kube-api-access-7zx8x") pod "56de51ce-dc44-4bda-9b13-191c2cddae41" (UID: "56de51ce-dc44-4bda-9b13-191c2cddae41"). InnerVolumeSpecName "kube-api-access-7zx8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:04 crc kubenswrapper[4706]: I1127 07:30:04.972538 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" (UID: "f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.056488 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zx8x\" (UniqueName: \"kubernetes.io/projected/56de51ce-dc44-4bda-9b13-191c2cddae41-kube-api-access-7zx8x\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.056512 4706 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/56de51ce-dc44-4bda-9b13-191c2cddae41-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.056523 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4jcs\" (UniqueName: \"kubernetes.io/projected/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-kube-api-access-b4jcs\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.056531 4706 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.473975 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" event={"ID":"56de51ce-dc44-4bda-9b13-191c2cddae41","Type":"ContainerDied","Data":"a25c5c95324cbb513658fd6f3c56438bb3c01297c81f62f9569cf0a9075bd385"} Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.475790 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a25c5c95324cbb513658fd6f3c56438bb3c01297c81f62f9569cf0a9075bd385" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.474022 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.477034 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" event={"ID":"f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681","Type":"ContainerDied","Data":"059ccd2b1036af6c8baf2cf159988e4f82ef434c66c3e0e9481690d3a9e5d0cb"} Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.477109 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="059ccd2b1036af6c8baf2cf159988e4f82ef434c66c3e0e9481690d3a9e5d0cb" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.477435 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.478861 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.478928 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.478951 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:05 crc kubenswrapper[4706]: I1127 07:30:05.478967 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:07 crc kubenswrapper[4706]: I1127 07:30:07.424194 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:07 crc kubenswrapper[4706]: I1127 07:30:07.494309 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:30:07 crc kubenswrapper[4706]: I1127 07:30:07.503129 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:07 crc kubenswrapper[4706]: I1127 07:30:07.503188 4706 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 07:30:07 crc kubenswrapper[4706]: I1127 07:30:07.509093 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:07 crc kubenswrapper[4706]: I1127 07:30:07.516689 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:08 crc kubenswrapper[4706]: I1127 07:30:08.919488 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 27 07:30:08 crc kubenswrapper[4706]: I1127 07:30:08.933108 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:30:09 crc kubenswrapper[4706]: I1127 07:30:09.514308 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-httpd" containerID="cri-o://87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98" gracePeriod=30 Nov 27 07:30:09 crc kubenswrapper[4706]: I1127 07:30:09.514767 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-log" containerID="cri-o://aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0" gracePeriod=30 Nov 27 07:30:10 crc kubenswrapper[4706]: I1127 07:30:10.522961 4706 generic.go:334] "Generic (PLEG): container finished" podID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerID="aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0" exitCode=143 Nov 27 07:30:10 crc kubenswrapper[4706]: I1127 07:30:10.523896 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-log" containerID="cri-o://b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1" gracePeriod=30 Nov 27 07:30:10 crc kubenswrapper[4706]: I1127 07:30:10.524304 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8913e171-ae2c-437c-850f-a84fa2918d5b","Type":"ContainerDied","Data":"aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0"} Nov 27 07:30:10 crc kubenswrapper[4706]: I1127 07:30:10.524670 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-httpd" containerID="cri-o://f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904" gracePeriod=30 Nov 27 07:30:11 crc kubenswrapper[4706]: I1127 07:30:11.534688 4706 generic.go:334] "Generic (PLEG): container finished" podID="26d60511-ef2e-4335-af43-816b2d2389fd" containerID="b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1" exitCode=143 Nov 27 07:30:11 crc kubenswrapper[4706]: I1127 07:30:11.534732 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26d60511-ef2e-4335-af43-816b2d2389fd","Type":"ContainerDied","Data":"b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1"} Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.059454 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.189851 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-config-data\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190239 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-iscsi\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190271 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-lib-modules\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190302 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-scripts\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190334 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-run\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190355 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190365 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190382 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-nvme\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190401 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-sys\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190449 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190473 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-run" (OuterVolumeSpecName: "run") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190500 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190527 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-httpd-run\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190509 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.190556 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-sys" (OuterVolumeSpecName: "sys") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191011 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191083 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-var-locks-brick\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191129 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq4br\" (UniqueName: \"kubernetes.io/projected/8913e171-ae2c-437c-850f-a84fa2918d5b-kube-api-access-nq4br\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191196 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191153 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-dev\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191329 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-logs\") pod \"8913e171-ae2c-437c-850f-a84fa2918d5b\" (UID: \"8913e171-ae2c-437c-850f-a84fa2918d5b\") " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191404 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-dev" (OuterVolumeSpecName: "dev") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191748 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191771 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191783 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191794 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191806 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191815 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191825 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191835 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8913e171-ae2c-437c-850f-a84fa2918d5b-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.191876 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-logs" (OuterVolumeSpecName: "logs") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.195929 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.196306 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.196917 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-scripts" (OuterVolumeSpecName: "scripts") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.198113 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8913e171-ae2c-437c-850f-a84fa2918d5b-kube-api-access-nq4br" (OuterVolumeSpecName: "kube-api-access-nq4br") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "kube-api-access-nq4br". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.234188 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-config-data" (OuterVolumeSpecName: "config-data") pod "8913e171-ae2c-437c-850f-a84fa2918d5b" (UID: "8913e171-ae2c-437c-850f-a84fa2918d5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.293426 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.293491 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.293511 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.293525 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq4br\" (UniqueName: \"kubernetes.io/projected/8913e171-ae2c-437c-850f-a84fa2918d5b-kube-api-access-nq4br\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.293538 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8913e171-ae2c-437c-850f-a84fa2918d5b-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.293549 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8913e171-ae2c-437c-850f-a84fa2918d5b-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.307728 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.310897 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.395134 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.395183 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.572014 4706 generic.go:334] "Generic (PLEG): container finished" podID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerID="87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98" exitCode=0 Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.572077 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8913e171-ae2c-437c-850f-a84fa2918d5b","Type":"ContainerDied","Data":"87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98"} Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.572144 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.572530 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8913e171-ae2c-437c-850f-a84fa2918d5b","Type":"ContainerDied","Data":"d868f6d216b800b5f18a8e8900551f2a8f1e6a846926acaabac96c2459e55ab8"} Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.572748 4706 scope.go:117] "RemoveContainer" containerID="87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.619909 4706 scope.go:117] "RemoveContainer" containerID="aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.633158 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.652852 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.665027 4706 scope.go:117] "RemoveContainer" containerID="87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98" Nov 27 07:30:13 crc kubenswrapper[4706]: E1127 07:30:13.665546 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98\": container with ID starting with 87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98 not found: ID does not exist" containerID="87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.665588 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98"} err="failed to get container status \"87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98\": rpc error: code = NotFound desc = could not find container \"87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98\": container with ID starting with 87c031c19fd11ad70fbff9a1564c0a013fc8cce2956b47d7349707b9bc37fd98 not found: ID does not exist" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.665615 4706 scope.go:117] "RemoveContainer" containerID="aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0" Nov 27 07:30:13 crc kubenswrapper[4706]: E1127 07:30:13.666124 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0\": container with ID starting with aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0 not found: ID does not exist" containerID="aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.666156 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0"} err="failed to get container status \"aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0\": rpc error: code = NotFound desc = could not find container \"aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0\": container with ID starting with aa253c6c4b744cfb270ab8d054a5c5bb0b22645a3a5ddee98dd857e0086dffd0 not found: ID does not exist" Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.789945 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp"] Nov 27 07:30:13 crc kubenswrapper[4706]: I1127 07:30:13.795003 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-2-cleaner-29403810-nr9vp"] Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.029740 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109078 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-scripts\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109122 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-run\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109173 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109251 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kvx7\" (UniqueName: \"kubernetes.io/projected/26d60511-ef2e-4335-af43-816b2d2389fd-kube-api-access-7kvx7\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109293 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-sys\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109317 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-nvme\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109336 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-iscsi\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109373 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-logs\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109394 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-lib-modules\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109435 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-dev\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109460 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109494 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-var-locks-brick\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109515 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-config-data\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109561 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-httpd-run\") pod \"26d60511-ef2e-4335-af43-816b2d2389fd\" (UID: \"26d60511-ef2e-4335-af43-816b2d2389fd\") " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109894 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.109994 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-dev" (OuterVolumeSpecName: "dev") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.110001 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-sys" (OuterVolumeSpecName: "sys") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.110046 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.110013 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.110084 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.110130 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.110134 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-run" (OuterVolumeSpecName: "run") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.110235 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-logs" (OuterVolumeSpecName: "logs") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.114579 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance-cache") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.114856 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-scripts" (OuterVolumeSpecName: "scripts") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.117030 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26d60511-ef2e-4335-af43-816b2d2389fd-kube-api-access-7kvx7" (OuterVolumeSpecName: "kube-api-access-7kvx7") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "kube-api-access-7kvx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.117731 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.151452 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-config-data" (OuterVolumeSpecName: "config-data") pod "26d60511-ef2e-4335-af43-816b2d2389fd" (UID: "26d60511-ef2e-4335-af43-816b2d2389fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210677 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210708 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210749 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210762 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kvx7\" (UniqueName: \"kubernetes.io/projected/26d60511-ef2e-4335-af43-816b2d2389fd-kube-api-access-7kvx7\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210772 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210780 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210788 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210798 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210805 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210815 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210828 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210836 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26d60511-ef2e-4335-af43-816b2d2389fd-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210844 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d60511-ef2e-4335-af43-816b2d2389fd-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.210852 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26d60511-ef2e-4335-af43-816b2d2389fd-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.223393 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.227873 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.312897 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.312930 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.584555 4706 generic.go:334] "Generic (PLEG): container finished" podID="26d60511-ef2e-4335-af43-816b2d2389fd" containerID="f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904" exitCode=0 Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.584631 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26d60511-ef2e-4335-af43-816b2d2389fd","Type":"ContainerDied","Data":"f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904"} Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.584661 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26d60511-ef2e-4335-af43-816b2d2389fd","Type":"ContainerDied","Data":"f6f2203b0cfbec31580b065a9633cf67c158a40cad0dffb4ace1ac65fd3318ee"} Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.584654 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.584703 4706 scope.go:117] "RemoveContainer" containerID="f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.621729 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.628305 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.634267 4706 scope.go:117] "RemoveContainer" containerID="b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.667459 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6"] Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.677197 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29403810-zgrl6"] Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.678424 4706 scope.go:117] "RemoveContainer" containerID="f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904" Nov 27 07:30:14 crc kubenswrapper[4706]: E1127 07:30:14.678985 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904\": container with ID starting with f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904 not found: ID does not exist" containerID="f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.679100 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904"} err="failed to get container status \"f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904\": rpc error: code = NotFound desc = could not find container \"f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904\": container with ID starting with f97b7cb707de04ff128918279d62846a002394b5135e10d2bd6e124efc37c904 not found: ID does not exist" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.679199 4706 scope.go:117] "RemoveContainer" containerID="b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1" Nov 27 07:30:14 crc kubenswrapper[4706]: E1127 07:30:14.681056 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1\": container with ID starting with b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1 not found: ID does not exist" containerID="b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.681086 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1"} err="failed to get container status \"b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1\": rpc error: code = NotFound desc = could not find container \"b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1\": container with ID starting with b435830524def2ac29951769e1fc9e95efbf2e041af2c0246ceebb49e1ae61a1 not found: ID does not exist" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.785124 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" path="/var/lib/kubelet/pods/26d60511-ef2e-4335-af43-816b2d2389fd/volumes" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.785712 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56de51ce-dc44-4bda-9b13-191c2cddae41" path="/var/lib/kubelet/pods/56de51ce-dc44-4bda-9b13-191c2cddae41/volumes" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.786343 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" path="/var/lib/kubelet/pods/8913e171-ae2c-437c-850f-a84fa2918d5b/volumes" Nov 27 07:30:14 crc kubenswrapper[4706]: I1127 07:30:14.787295 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" path="/var/lib/kubelet/pods/f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681/volumes" Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.177270 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.177317 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.177352 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.178358 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6fa6ce74ad58cab1c7d76467a227f52ada0afc016b50f3d7c2b5893e9773b0ee"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.178407 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://6fa6ce74ad58cab1c7d76467a227f52ada0afc016b50f3d7c2b5893e9773b0ee" gracePeriod=600 Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.194332 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.194598 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-log" containerID="cri-o://ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d" gracePeriod=30 Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.194723 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-httpd" containerID="cri-o://5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb" gracePeriod=30 Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.595991 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="6fa6ce74ad58cab1c7d76467a227f52ada0afc016b50f3d7c2b5893e9773b0ee" exitCode=0 Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.596067 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"6fa6ce74ad58cab1c7d76467a227f52ada0afc016b50f3d7c2b5893e9773b0ee"} Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.596676 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900"} Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.596765 4706 scope.go:117] "RemoveContainer" containerID="d7703c2342c25e6f9c346c6d641fa2f157694340cd2c97e78820343e2f8c3744" Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.599578 4706 generic.go:334] "Generic (PLEG): container finished" podID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerID="ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d" exitCode=143 Nov 27 07:30:15 crc kubenswrapper[4706]: I1127 07:30:15.599771 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a5ff227b-1215-462b-824d-9081fcfb9fff","Type":"ContainerDied","Data":"ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d"} Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.594634 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.635730 4706 generic.go:334] "Generic (PLEG): container finished" podID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerID="5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb" exitCode=0 Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.635791 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a5ff227b-1215-462b-824d-9081fcfb9fff","Type":"ContainerDied","Data":"5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb"} Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.635814 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.636251 4706 scope.go:117] "RemoveContainer" containerID="5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.636234 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a5ff227b-1215-462b-824d-9081fcfb9fff","Type":"ContainerDied","Data":"b3ad6108d735a77dc7f9dc27a201b8d466406b6548ee4b8424b1ba8d6ef8d1db"} Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.662154 4706 scope.go:117] "RemoveContainer" containerID="ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681684 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-config-data\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681759 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6sxm\" (UniqueName: \"kubernetes.io/projected/a5ff227b-1215-462b-824d-9081fcfb9fff-kube-api-access-m6sxm\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681788 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681821 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-httpd-run\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681847 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-sys\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681881 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-iscsi\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681904 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681926 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-var-locks-brick\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681950 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-dev\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681959 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.681974 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-run\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682007 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-run" (OuterVolumeSpecName: "run") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682045 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-scripts\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682068 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-lib-modules\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682049 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-dev" (OuterVolumeSpecName: "dev") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682070 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-sys" (OuterVolumeSpecName: "sys") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682124 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-logs\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682057 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682146 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682153 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-nvme\") pod \"a5ff227b-1215-462b-824d-9081fcfb9fff\" (UID: \"a5ff227b-1215-462b-824d-9081fcfb9fff\") " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682177 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682487 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682643 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682662 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682674 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682684 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682696 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682706 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682717 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682726 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a5ff227b-1215-462b-824d-9081fcfb9fff-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682827 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-logs" (OuterVolumeSpecName: "logs") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.682920 4706 scope.go:117] "RemoveContainer" containerID="5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb" Nov 27 07:30:18 crc kubenswrapper[4706]: E1127 07:30:18.684278 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb\": container with ID starting with 5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb not found: ID does not exist" containerID="5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.684315 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb"} err="failed to get container status \"5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb\": rpc error: code = NotFound desc = could not find container \"5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb\": container with ID starting with 5b9cbe5d1a67b256df0a62de0e712cc58137578ffafe414055a5439fe42570fb not found: ID does not exist" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.684339 4706 scope.go:117] "RemoveContainer" containerID="ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d" Nov 27 07:30:18 crc kubenswrapper[4706]: E1127 07:30:18.684717 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d\": container with ID starting with ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d not found: ID does not exist" containerID="ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.684751 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d"} err="failed to get container status \"ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d\": rpc error: code = NotFound desc = could not find container \"ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d\": container with ID starting with ab769075bf76070a6232534042da238eb7c8012b85c2d404ad589df9e5203a2d not found: ID does not exist" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.687590 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.688344 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ff227b-1215-462b-824d-9081fcfb9fff-kube-api-access-m6sxm" (OuterVolumeSpecName: "kube-api-access-m6sxm") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "kube-api-access-m6sxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.690747 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.698750 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-scripts" (OuterVolumeSpecName: "scripts") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.719388 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-config-data" (OuterVolumeSpecName: "config-data") pod "a5ff227b-1215-462b-824d-9081fcfb9fff" (UID: "a5ff227b-1215-462b-824d-9081fcfb9fff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.785040 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6sxm\" (UniqueName: \"kubernetes.io/projected/a5ff227b-1215-462b-824d-9081fcfb9fff-kube-api-access-m6sxm\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.785112 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.785131 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.785143 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.785153 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5ff227b-1215-462b-824d-9081fcfb9fff-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.785180 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ff227b-1215-462b-824d-9081fcfb9fff-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.799377 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.808753 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.887002 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.887032 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.964187 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:30:18 crc kubenswrapper[4706]: I1127 07:30:18.974933 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.225181 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr"] Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.242869 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29403810-xg8rr"] Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.593968 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d92kp"] Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.604519 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-d92kp"] Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.649882 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance31fa-account-delete-kxxtl"] Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650397 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56de51ce-dc44-4bda-9b13-191c2cddae41" containerName="glance-cache-glance-default-single-2-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650428 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="56de51ce-dc44-4bda-9b13-191c2cddae41" containerName="glance-cache-glance-default-single-2-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650454 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650463 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650476 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650483 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650496 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650505 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650519 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650527 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650536 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f94cc7-9f7a-4af5-a9ce-a403adabf163" containerName="glance-cache-glance-default-single-0-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650543 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f94cc7-9f7a-4af5-a9ce-a403adabf163" containerName="glance-cache-glance-default-single-0-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650562 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8" containerName="collect-profiles" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650569 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8" containerName="collect-profiles" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650591 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650600 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650612 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" containerName="glance-cache-glance-default-single-1-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650621 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" containerName="glance-cache-glance-default-single-1-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: E1127 07:30:19.650630 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650637 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650774 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650784 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="05f94cc7-9f7a-4af5-a9ce-a403adabf163" containerName="glance-cache-glance-default-single-0-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650791 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650801 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650814 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8913e171-ae2c-437c-850f-a84fa2918d5b" containerName="glance-httpd" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650823 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a3c7ba-d205-40f7-b2f6-97a3fc2fd681" containerName="glance-cache-glance-default-single-1-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650835 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650846 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="26d60511-ef2e-4335-af43-816b2d2389fd" containerName="glance-log" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650856 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c74f2b-bf48-46fb-9071-e0cd00e4f5a8" containerName="collect-profiles" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.650867 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="56de51ce-dc44-4bda-9b13-191c2cddae41" containerName="glance-cache-glance-default-single-2-cleaner" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.651663 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.668448 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance31fa-account-delete-kxxtl"] Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.807775 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxbck\" (UniqueName: \"kubernetes.io/projected/0a78978a-dc02-416d-8f7f-0320cd606588-kube-api-access-gxbck\") pod \"glance31fa-account-delete-kxxtl\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.807861 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a78978a-dc02-416d-8f7f-0320cd606588-operator-scripts\") pod \"glance31fa-account-delete-kxxtl\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.909474 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxbck\" (UniqueName: \"kubernetes.io/projected/0a78978a-dc02-416d-8f7f-0320cd606588-kube-api-access-gxbck\") pod \"glance31fa-account-delete-kxxtl\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.909583 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a78978a-dc02-416d-8f7f-0320cd606588-operator-scripts\") pod \"glance31fa-account-delete-kxxtl\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.910347 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a78978a-dc02-416d-8f7f-0320cd606588-operator-scripts\") pod \"glance31fa-account-delete-kxxtl\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.940639 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxbck\" (UniqueName: \"kubernetes.io/projected/0a78978a-dc02-416d-8f7f-0320cd606588-kube-api-access-gxbck\") pod \"glance31fa-account-delete-kxxtl\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:19 crc kubenswrapper[4706]: I1127 07:30:19.976396 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:20 crc kubenswrapper[4706]: I1127 07:30:20.215530 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance31fa-account-delete-kxxtl"] Nov 27 07:30:20 crc kubenswrapper[4706]: W1127 07:30:20.220030 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a78978a_dc02_416d_8f7f_0320cd606588.slice/crio-5fd20cd9dd09ed31b97a6f29a54e59ce98afe57c2bd2b24765a2fb13c4c857dd WatchSource:0}: Error finding container 5fd20cd9dd09ed31b97a6f29a54e59ce98afe57c2bd2b24765a2fb13c4c857dd: Status 404 returned error can't find the container with id 5fd20cd9dd09ed31b97a6f29a54e59ce98afe57c2bd2b24765a2fb13c4c857dd Nov 27 07:30:20 crc kubenswrapper[4706]: I1127 07:30:20.675334 4706 generic.go:334] "Generic (PLEG): container finished" podID="0a78978a-dc02-416d-8f7f-0320cd606588" containerID="bd810a5801a6e24867ccfb8c4a0932f5ae2edea4a3acb69e4922725d53f1114d" exitCode=0 Nov 27 07:30:20 crc kubenswrapper[4706]: I1127 07:30:20.675436 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" event={"ID":"0a78978a-dc02-416d-8f7f-0320cd606588","Type":"ContainerDied","Data":"bd810a5801a6e24867ccfb8c4a0932f5ae2edea4a3acb69e4922725d53f1114d"} Nov 27 07:30:20 crc kubenswrapper[4706]: I1127 07:30:20.675607 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" event={"ID":"0a78978a-dc02-416d-8f7f-0320cd606588","Type":"ContainerStarted","Data":"5fd20cd9dd09ed31b97a6f29a54e59ce98afe57c2bd2b24765a2fb13c4c857dd"} Nov 27 07:30:20 crc kubenswrapper[4706]: I1127 07:30:20.785457 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05f94cc7-9f7a-4af5-a9ce-a403adabf163" path="/var/lib/kubelet/pods/05f94cc7-9f7a-4af5-a9ce-a403adabf163/volumes" Nov 27 07:30:20 crc kubenswrapper[4706]: I1127 07:30:20.786330 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="180c49cf-2510-48db-b203-523b3933d5a0" path="/var/lib/kubelet/pods/180c49cf-2510-48db-b203-523b3933d5a0/volumes" Nov 27 07:30:20 crc kubenswrapper[4706]: I1127 07:30:20.787120 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5ff227b-1215-462b-824d-9081fcfb9fff" path="/var/lib/kubelet/pods/a5ff227b-1215-462b-824d-9081fcfb9fff/volumes" Nov 27 07:30:21 crc kubenswrapper[4706]: I1127 07:30:21.973808 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.040243 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxbck\" (UniqueName: \"kubernetes.io/projected/0a78978a-dc02-416d-8f7f-0320cd606588-kube-api-access-gxbck\") pod \"0a78978a-dc02-416d-8f7f-0320cd606588\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.040352 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a78978a-dc02-416d-8f7f-0320cd606588-operator-scripts\") pod \"0a78978a-dc02-416d-8f7f-0320cd606588\" (UID: \"0a78978a-dc02-416d-8f7f-0320cd606588\") " Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.041160 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a78978a-dc02-416d-8f7f-0320cd606588-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a78978a-dc02-416d-8f7f-0320cd606588" (UID: "0a78978a-dc02-416d-8f7f-0320cd606588"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.063426 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a78978a-dc02-416d-8f7f-0320cd606588-kube-api-access-gxbck" (OuterVolumeSpecName: "kube-api-access-gxbck") pod "0a78978a-dc02-416d-8f7f-0320cd606588" (UID: "0a78978a-dc02-416d-8f7f-0320cd606588"). InnerVolumeSpecName "kube-api-access-gxbck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.144608 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a78978a-dc02-416d-8f7f-0320cd606588-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.144738 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxbck\" (UniqueName: \"kubernetes.io/projected/0a78978a-dc02-416d-8f7f-0320cd606588-kube-api-access-gxbck\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.701664 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" event={"ID":"0a78978a-dc02-416d-8f7f-0320cd606588","Type":"ContainerDied","Data":"5fd20cd9dd09ed31b97a6f29a54e59ce98afe57c2bd2b24765a2fb13c4c857dd"} Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.701866 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fd20cd9dd09ed31b97a6f29a54e59ce98afe57c2bd2b24765a2fb13c4c857dd" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.701769 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance31fa-account-delete-kxxtl" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.914334 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:30:22 crc kubenswrapper[4706]: E1127 07:30:22.914605 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a78978a-dc02-416d-8f7f-0320cd606588" containerName="mariadb-account-delete" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.914621 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a78978a-dc02-416d-8f7f-0320cd606588" containerName="mariadb-account-delete" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.914728 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a78978a-dc02-416d-8f7f-0320cd606588" containerName="mariadb-account-delete" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.915133 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.916668 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-trwjt" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.918443 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.919278 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.919338 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 27 07:30:22 crc kubenswrapper[4706]: I1127 07:30:22.923003 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.058644 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hwrn\" (UniqueName: \"kubernetes.io/projected/2482c5b8-e170-4a54-9505-d422bc31d390-kube-api-access-8hwrn\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.058714 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.059035 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-scripts\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.059104 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.160484 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-scripts\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.160535 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.160596 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hwrn\" (UniqueName: \"kubernetes.io/projected/2482c5b8-e170-4a54-9505-d422bc31d390-kube-api-access-8hwrn\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.160621 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.161506 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.162170 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-scripts\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.165742 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.176843 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hwrn\" (UniqueName: \"kubernetes.io/projected/2482c5b8-e170-4a54-9505-d422bc31d390-kube-api-access-8hwrn\") pod \"openstackclient\" (UID: \"2482c5b8-e170-4a54-9505-d422bc31d390\") " pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.229594 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.668239 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 27 07:30:23 crc kubenswrapper[4706]: I1127 07:30:23.710031 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"2482c5b8-e170-4a54-9505-d422bc31d390","Type":"ContainerStarted","Data":"d1a1c5e4c319e2427668b73f24bee7fa4d03141ee557ff9c1e91621fb87aabef"} Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.684037 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-9x2v2"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.695637 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-9x2v2"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.704335 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance31fa-account-delete-kxxtl"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.711859 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-31fa-account-create-update-79nsh"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.718285 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance31fa-account-delete-kxxtl"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.721497 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"2482c5b8-e170-4a54-9505-d422bc31d390","Type":"ContainerStarted","Data":"6a03e2782beb14901cd858d5d7c7030c714b349c84f35ecc8373b6d735daa072"} Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.725064 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-31fa-account-create-update-79nsh"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.739074 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.73905487 podStartE2EDuration="2.73905487s" podCreationTimestamp="2025-11-27 07:30:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:24.738086244 +0000 UTC m=+1308.627677064" watchObservedRunningTime="2025-11-27 07:30:24.73905487 +0000 UTC m=+1308.628645680" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.770287 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-st6x5"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.771551 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.785592 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a78978a-dc02-416d-8f7f-0320cd606588" path="/var/lib/kubelet/pods/0a78978a-dc02-416d-8f7f-0320cd606588/volumes" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.798418 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b4fed3d-33e8-4bdc-a096-1157a74ae784" path="/var/lib/kubelet/pods/5b4fed3d-33e8-4bdc-a096-1157a74ae784/volumes" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.800529 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec594dd5-ee4d-4587-85ca-244798209a0e" path="/var/lib/kubelet/pods/ec594dd5-ee4d-4587-85ca-244798209a0e/volumes" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.801207 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-st6x5"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.863401 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-6f10-account-create-update-zmjz6"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.864175 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.865914 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.874954 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6f10-account-create-update-zmjz6"] Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.884060 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf92f\" (UniqueName: \"kubernetes.io/projected/5f704bc5-b657-4454-b835-18c278b0bb24-kube-api-access-pf92f\") pod \"glance-db-create-st6x5\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.884104 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f704bc5-b657-4454-b835-18c278b0bb24-operator-scripts\") pod \"glance-db-create-st6x5\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.984962 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-operator-scripts\") pod \"glance-6f10-account-create-update-zmjz6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.985075 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf92f\" (UniqueName: \"kubernetes.io/projected/5f704bc5-b657-4454-b835-18c278b0bb24-kube-api-access-pf92f\") pod \"glance-db-create-st6x5\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.985096 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f704bc5-b657-4454-b835-18c278b0bb24-operator-scripts\") pod \"glance-db-create-st6x5\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.985147 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq65f\" (UniqueName: \"kubernetes.io/projected/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-kube-api-access-sq65f\") pod \"glance-6f10-account-create-update-zmjz6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:24 crc kubenswrapper[4706]: I1127 07:30:24.985978 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f704bc5-b657-4454-b835-18c278b0bb24-operator-scripts\") pod \"glance-db-create-st6x5\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.009958 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf92f\" (UniqueName: \"kubernetes.io/projected/5f704bc5-b657-4454-b835-18c278b0bb24-kube-api-access-pf92f\") pod \"glance-db-create-st6x5\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.087016 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq65f\" (UniqueName: \"kubernetes.io/projected/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-kube-api-access-sq65f\") pod \"glance-6f10-account-create-update-zmjz6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.087152 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-operator-scripts\") pod \"glance-6f10-account-create-update-zmjz6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.087720 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.087903 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-operator-scripts\") pod \"glance-6f10-account-create-update-zmjz6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.109108 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq65f\" (UniqueName: \"kubernetes.io/projected/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-kube-api-access-sq65f\") pod \"glance-6f10-account-create-update-zmjz6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.179266 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.525469 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-st6x5"] Nov 27 07:30:25 crc kubenswrapper[4706]: W1127 07:30:25.532977 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f704bc5_b657_4454_b835_18c278b0bb24.slice/crio-b96dac6d0d0c291768a97780a06c5ccb06879f544e09236b71a2f4ad0129fb21 WatchSource:0}: Error finding container b96dac6d0d0c291768a97780a06c5ccb06879f544e09236b71a2f4ad0129fb21: Status 404 returned error can't find the container with id b96dac6d0d0c291768a97780a06c5ccb06879f544e09236b71a2f4ad0129fb21 Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.658716 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6f10-account-create-update-zmjz6"] Nov 27 07:30:25 crc kubenswrapper[4706]: W1127 07:30:25.663663 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96793f14_42e4_4a6f_8e18_cefa7e3f4bb6.slice/crio-6dd335793692d6b73de9fee536a08429193cf6507fc4e0d0fbe17830bfddfed0 WatchSource:0}: Error finding container 6dd335793692d6b73de9fee536a08429193cf6507fc4e0d0fbe17830bfddfed0: Status 404 returned error can't find the container with id 6dd335793692d6b73de9fee536a08429193cf6507fc4e0d0fbe17830bfddfed0 Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.729349 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" event={"ID":"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6","Type":"ContainerStarted","Data":"6dd335793692d6b73de9fee536a08429193cf6507fc4e0d0fbe17830bfddfed0"} Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.732215 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-st6x5" event={"ID":"5f704bc5-b657-4454-b835-18c278b0bb24","Type":"ContainerStarted","Data":"900f66aebd0ca0ecf88ac5851201956fb93ca9d8a5fd8e9f071f080eb3ce0876"} Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.732290 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-st6x5" event={"ID":"5f704bc5-b657-4454-b835-18c278b0bb24","Type":"ContainerStarted","Data":"b96dac6d0d0c291768a97780a06c5ccb06879f544e09236b71a2f4ad0129fb21"} Nov 27 07:30:25 crc kubenswrapper[4706]: I1127 07:30:25.747060 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-st6x5" podStartSLOduration=1.747041117 podStartE2EDuration="1.747041117s" podCreationTimestamp="2025-11-27 07:30:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:25.744420461 +0000 UTC m=+1309.634011271" watchObservedRunningTime="2025-11-27 07:30:25.747041117 +0000 UTC m=+1309.636631927" Nov 27 07:30:26 crc kubenswrapper[4706]: I1127 07:30:26.741119 4706 generic.go:334] "Generic (PLEG): container finished" podID="5f704bc5-b657-4454-b835-18c278b0bb24" containerID="900f66aebd0ca0ecf88ac5851201956fb93ca9d8a5fd8e9f071f080eb3ce0876" exitCode=0 Nov 27 07:30:26 crc kubenswrapper[4706]: I1127 07:30:26.741190 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-st6x5" event={"ID":"5f704bc5-b657-4454-b835-18c278b0bb24","Type":"ContainerDied","Data":"900f66aebd0ca0ecf88ac5851201956fb93ca9d8a5fd8e9f071f080eb3ce0876"} Nov 27 07:30:26 crc kubenswrapper[4706]: I1127 07:30:26.744276 4706 generic.go:334] "Generic (PLEG): container finished" podID="96793f14-42e4-4a6f-8e18-cefa7e3f4bb6" containerID="1e657a803e73c8309a59300617776cc8169e2d9c54cf6150e4e21e636ef8156c" exitCode=0 Nov 27 07:30:26 crc kubenswrapper[4706]: I1127 07:30:26.744328 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" event={"ID":"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6","Type":"ContainerDied","Data":"1e657a803e73c8309a59300617776cc8169e2d9c54cf6150e4e21e636ef8156c"} Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.163378 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.168568 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.257196 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-operator-scripts\") pod \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.257243 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq65f\" (UniqueName: \"kubernetes.io/projected/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-kube-api-access-sq65f\") pod \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\" (UID: \"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6\") " Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.257295 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f704bc5-b657-4454-b835-18c278b0bb24-operator-scripts\") pod \"5f704bc5-b657-4454-b835-18c278b0bb24\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.257356 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf92f\" (UniqueName: \"kubernetes.io/projected/5f704bc5-b657-4454-b835-18c278b0bb24-kube-api-access-pf92f\") pod \"5f704bc5-b657-4454-b835-18c278b0bb24\" (UID: \"5f704bc5-b657-4454-b835-18c278b0bb24\") " Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.257987 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "96793f14-42e4-4a6f-8e18-cefa7e3f4bb6" (UID: "96793f14-42e4-4a6f-8e18-cefa7e3f4bb6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.257989 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f704bc5-b657-4454-b835-18c278b0bb24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f704bc5-b657-4454-b835-18c278b0bb24" (UID: "5f704bc5-b657-4454-b835-18c278b0bb24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.263274 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f704bc5-b657-4454-b835-18c278b0bb24-kube-api-access-pf92f" (OuterVolumeSpecName: "kube-api-access-pf92f") pod "5f704bc5-b657-4454-b835-18c278b0bb24" (UID: "5f704bc5-b657-4454-b835-18c278b0bb24"). InnerVolumeSpecName "kube-api-access-pf92f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.263508 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-kube-api-access-sq65f" (OuterVolumeSpecName: "kube-api-access-sq65f") pod "96793f14-42e4-4a6f-8e18-cefa7e3f4bb6" (UID: "96793f14-42e4-4a6f-8e18-cefa7e3f4bb6"). InnerVolumeSpecName "kube-api-access-sq65f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.359490 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq65f\" (UniqueName: \"kubernetes.io/projected/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-kube-api-access-sq65f\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.359523 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.359534 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f704bc5-b657-4454-b835-18c278b0bb24-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.359543 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf92f\" (UniqueName: \"kubernetes.io/projected/5f704bc5-b657-4454-b835-18c278b0bb24-kube-api-access-pf92f\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.771195 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" event={"ID":"96793f14-42e4-4a6f-8e18-cefa7e3f4bb6","Type":"ContainerDied","Data":"6dd335793692d6b73de9fee536a08429193cf6507fc4e0d0fbe17830bfddfed0"} Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.771299 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6f10-account-create-update-zmjz6" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.771310 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dd335793692d6b73de9fee536a08429193cf6507fc4e0d0fbe17830bfddfed0" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.772974 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-st6x5" event={"ID":"5f704bc5-b657-4454-b835-18c278b0bb24","Type":"ContainerDied","Data":"b96dac6d0d0c291768a97780a06c5ccb06879f544e09236b71a2f4ad0129fb21"} Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.773014 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b96dac6d0d0c291768a97780a06c5ccb06879f544e09236b71a2f4ad0129fb21" Nov 27 07:30:28 crc kubenswrapper[4706]: I1127 07:30:28.773169 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-st6x5" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.092538 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-cskxq"] Nov 27 07:30:30 crc kubenswrapper[4706]: E1127 07:30:30.095595 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96793f14-42e4-4a6f-8e18-cefa7e3f4bb6" containerName="mariadb-account-create-update" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.095638 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="96793f14-42e4-4a6f-8e18-cefa7e3f4bb6" containerName="mariadb-account-create-update" Nov 27 07:30:30 crc kubenswrapper[4706]: E1127 07:30:30.095725 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f704bc5-b657-4454-b835-18c278b0bb24" containerName="mariadb-database-create" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.095747 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f704bc5-b657-4454-b835-18c278b0bb24" containerName="mariadb-database-create" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.096471 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="96793f14-42e4-4a6f-8e18-cefa7e3f4bb6" containerName="mariadb-account-create-update" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.096520 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f704bc5-b657-4454-b835-18c278b0bb24" containerName="mariadb-database-create" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.097746 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.109155 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.109192 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-rfgtj" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.120157 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cskxq"] Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.184321 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-db-sync-config-data\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.184401 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-config-data\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.184460 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c44kr\" (UniqueName: \"kubernetes.io/projected/d8b675cb-a460-4c74-81bf-80223ae396f2-kube-api-access-c44kr\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.285552 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-db-sync-config-data\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.285604 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-config-data\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.285636 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c44kr\" (UniqueName: \"kubernetes.io/projected/d8b675cb-a460-4c74-81bf-80223ae396f2-kube-api-access-c44kr\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.290062 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-config-data\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.290492 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-db-sync-config-data\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.303140 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c44kr\" (UniqueName: \"kubernetes.io/projected/d8b675cb-a460-4c74-81bf-80223ae396f2-kube-api-access-c44kr\") pod \"glance-db-sync-cskxq\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.488870 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:30 crc kubenswrapper[4706]: W1127 07:30:30.963285 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8b675cb_a460_4c74_81bf_80223ae396f2.slice/crio-673f0be65e873408c386c6f57d4cc9ad8919c3148a97fbf210dc45cb84ab34b3 WatchSource:0}: Error finding container 673f0be65e873408c386c6f57d4cc9ad8919c3148a97fbf210dc45cb84ab34b3: Status 404 returned error can't find the container with id 673f0be65e873408c386c6f57d4cc9ad8919c3148a97fbf210dc45cb84ab34b3 Nov 27 07:30:30 crc kubenswrapper[4706]: I1127 07:30:30.969076 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cskxq"] Nov 27 07:30:31 crc kubenswrapper[4706]: I1127 07:30:31.797563 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cskxq" event={"ID":"d8b675cb-a460-4c74-81bf-80223ae396f2","Type":"ContainerStarted","Data":"96b638f010819e9ff57f884e5715c34289491bbb2242e29c18262152b06f7621"} Nov 27 07:30:31 crc kubenswrapper[4706]: I1127 07:30:31.798007 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cskxq" event={"ID":"d8b675cb-a460-4c74-81bf-80223ae396f2","Type":"ContainerStarted","Data":"673f0be65e873408c386c6f57d4cc9ad8919c3148a97fbf210dc45cb84ab34b3"} Nov 27 07:30:31 crc kubenswrapper[4706]: I1127 07:30:31.823119 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-cskxq" podStartSLOduration=1.823099311 podStartE2EDuration="1.823099311s" podCreationTimestamp="2025-11-27 07:30:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:31.816786799 +0000 UTC m=+1315.706377639" watchObservedRunningTime="2025-11-27 07:30:31.823099311 +0000 UTC m=+1315.712690121" Nov 27 07:30:34 crc kubenswrapper[4706]: I1127 07:30:34.828007 4706 generic.go:334] "Generic (PLEG): container finished" podID="d8b675cb-a460-4c74-81bf-80223ae396f2" containerID="96b638f010819e9ff57f884e5715c34289491bbb2242e29c18262152b06f7621" exitCode=0 Nov 27 07:30:34 crc kubenswrapper[4706]: I1127 07:30:34.828146 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cskxq" event={"ID":"d8b675cb-a460-4c74-81bf-80223ae396f2","Type":"ContainerDied","Data":"96b638f010819e9ff57f884e5715c34289491bbb2242e29c18262152b06f7621"} Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.283832 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.383345 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c44kr\" (UniqueName: \"kubernetes.io/projected/d8b675cb-a460-4c74-81bf-80223ae396f2-kube-api-access-c44kr\") pod \"d8b675cb-a460-4c74-81bf-80223ae396f2\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.383416 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-config-data\") pod \"d8b675cb-a460-4c74-81bf-80223ae396f2\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.383613 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-db-sync-config-data\") pod \"d8b675cb-a460-4c74-81bf-80223ae396f2\" (UID: \"d8b675cb-a460-4c74-81bf-80223ae396f2\") " Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.389925 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8b675cb-a460-4c74-81bf-80223ae396f2-kube-api-access-c44kr" (OuterVolumeSpecName: "kube-api-access-c44kr") pod "d8b675cb-a460-4c74-81bf-80223ae396f2" (UID: "d8b675cb-a460-4c74-81bf-80223ae396f2"). InnerVolumeSpecName "kube-api-access-c44kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.399553 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d8b675cb-a460-4c74-81bf-80223ae396f2" (UID: "d8b675cb-a460-4c74-81bf-80223ae396f2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.450542 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-config-data" (OuterVolumeSpecName: "config-data") pod "d8b675cb-a460-4c74-81bf-80223ae396f2" (UID: "d8b675cb-a460-4c74-81bf-80223ae396f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.485021 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c44kr\" (UniqueName: \"kubernetes.io/projected/d8b675cb-a460-4c74-81bf-80223ae396f2-kube-api-access-c44kr\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.485064 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.485074 4706 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8b675cb-a460-4c74-81bf-80223ae396f2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.853207 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cskxq" event={"ID":"d8b675cb-a460-4c74-81bf-80223ae396f2","Type":"ContainerDied","Data":"673f0be65e873408c386c6f57d4cc9ad8919c3148a97fbf210dc45cb84ab34b3"} Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.853287 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="673f0be65e873408c386c6f57d4cc9ad8919c3148a97fbf210dc45cb84ab34b3" Nov 27 07:30:36 crc kubenswrapper[4706]: I1127 07:30:36.853364 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cskxq" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.041518 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:30:38 crc kubenswrapper[4706]: E1127 07:30:38.043209 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b675cb-a460-4c74-81bf-80223ae396f2" containerName="glance-db-sync" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.043264 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b675cb-a460-4c74-81bf-80223ae396f2" containerName="glance-db-sync" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.043933 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b675cb-a460-4c74-81bf-80223ae396f2" containerName="glance-db-sync" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.047628 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.078906 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.081504 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-rfgtj" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.082880 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.088210 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.108492 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.151448 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.152726 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.168316 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.209930 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.209995 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7mrc\" (UniqueName: \"kubernetes.io/projected/5dfa1095-7cae-4bdd-9615-d9a447134b6f-kube-api-access-s7mrc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210125 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210198 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-config-data\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210369 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-dev\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210440 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210501 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210537 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210579 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210634 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-sys\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210645 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210665 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-logs\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210730 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210759 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-run\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.210813 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-scripts\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.238544 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312443 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7mrc\" (UniqueName: \"kubernetes.io/projected/5dfa1095-7cae-4bdd-9615-d9a447134b6f-kube-api-access-s7mrc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312505 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-dev\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312533 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-sys\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312550 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312568 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-config-data\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312646 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjkmf\" (UniqueName: \"kubernetes.io/projected/b2f1f7fc-9f01-4487-aec2-c5a34d501890-kube-api-access-rjkmf\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312689 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-logs\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312717 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312757 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-dev\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312773 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312795 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-scripts\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312901 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312839 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-dev\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312960 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312990 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.312997 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313193 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313308 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-run\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313308 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313347 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313393 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-config-data\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313427 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313451 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313461 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313474 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313556 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-sys\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313597 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-logs\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313661 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313693 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-run\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313697 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-sys\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313763 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-scripts\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313816 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313915 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-logs\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.313958 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-run\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.314022 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.314028 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.317453 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-config-data\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.320357 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-scripts\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.339232 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7mrc\" (UniqueName: \"kubernetes.io/projected/5dfa1095-7cae-4bdd-9615-d9a447134b6f-kube-api-access-s7mrc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.341301 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.342462 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.346556 4706 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.351858 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-1\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.358318 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.359715 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.365829 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.384253 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.385207 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419620 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-dev\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419699 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419721 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-dev\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419729 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-sys\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419780 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-sys\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419815 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjkmf\" (UniqueName: \"kubernetes.io/projected/b2f1f7fc-9f01-4487-aec2-c5a34d501890-kube-api-access-rjkmf\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419817 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419843 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-logs\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.419966 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420006 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420056 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420094 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-scripts\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420143 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-logs\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420171 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420188 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420265 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-run\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420311 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-config-data\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420343 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420372 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.420387 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.421036 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.421123 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.421049 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-run\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.421256 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.421378 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.435460 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-scripts\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.435717 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-config-data\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.446249 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjkmf\" (UniqueName: \"kubernetes.io/projected/b2f1f7fc-9f01-4487-aec2-c5a34d501890-kube-api-access-rjkmf\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.447373 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.451552 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.469019 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.522881 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-dev\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.522951 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.522982 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-run\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523006 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-run\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523039 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523065 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523095 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-logs\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523134 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523161 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523198 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523260 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523290 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523325 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-dev\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523353 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-scripts\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523388 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523414 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523442 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523479 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnszx\" (UniqueName: \"kubernetes.io/projected/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-kube-api-access-jnszx\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523509 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523535 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523564 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-sys\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523594 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523624 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffsq5\" (UniqueName: \"kubernetes.io/projected/d3d092de-ab7b-4cd0-a939-20afa3d0d999-kube-api-access-ffsq5\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523650 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-config-data\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523681 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-sys\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523716 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-logs\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523759 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.523783 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625587 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-logs\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625669 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625707 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625745 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625797 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625819 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625847 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-dev\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625871 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-scripts\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625907 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625903 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625939 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-dev\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625942 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625973 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626000 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626028 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.625905 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626080 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626099 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626099 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnszx\" (UniqueName: \"kubernetes.io/projected/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-kube-api-access-jnszx\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626155 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626177 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626204 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-sys\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626254 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626284 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffsq5\" (UniqueName: \"kubernetes.io/projected/d3d092de-ab7b-4cd0-a939-20afa3d0d999-kube-api-access-ffsq5\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626287 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-logs\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626306 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-config-data\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626360 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626364 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-sys\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626400 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-sys\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626466 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626784 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-logs\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626836 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-sys\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626895 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.626481 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-logs\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627363 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627412 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-dev\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627446 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627483 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-run\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627508 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-run\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627545 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627578 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.627824 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.628201 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.628254 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.628357 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.628400 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-run\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.628431 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-run\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.628459 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-dev\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.630951 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-scripts\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.636386 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.637327 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.645901 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.648637 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-config-data\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.649795 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnszx\" (UniqueName: \"kubernetes.io/projected/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-kube-api-access-jnszx\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.658275 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffsq5\" (UniqueName: \"kubernetes.io/projected/d3d092de-ab7b-4cd0-a939-20afa3d0d999-kube-api-access-ffsq5\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.665691 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.665804 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.666065 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.667474 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.719848 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.740585 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.833641 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:30:38 crc kubenswrapper[4706]: I1127 07:30:38.889620 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"5dfa1095-7cae-4bdd-9615-d9a447134b6f","Type":"ContainerStarted","Data":"68da73b505ac9edbc564772d42ef87c2e02a2e18a78e430864b988e5db158080"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.010353 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.026891 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.139193 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:39 crc kubenswrapper[4706]: W1127 07:30:39.154809 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0f52bf2_bba2_46d6_b97a_3c36f800ae4d.slice/crio-0cbff9fc12174144cc719e60fbbe7b582434ca878c5ceab392476c3c38c86d30 WatchSource:0}: Error finding container 0cbff9fc12174144cc719e60fbbe7b582434ca878c5ceab392476c3c38c86d30: Status 404 returned error can't find the container with id 0cbff9fc12174144cc719e60fbbe7b582434ca878c5ceab392476c3c38c86d30 Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.293177 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.903365 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"5dfa1095-7cae-4bdd-9615-d9a447134b6f","Type":"ContainerStarted","Data":"8967b07927196e003fbc8d9d9c7b1573d12469c74464c3207c0b7aa091a96cbf"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.904200 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"5dfa1095-7cae-4bdd-9615-d9a447134b6f","Type":"ContainerStarted","Data":"dd37542636329515e16d92d078921902e2eb76643303dd071ecb13da84e7438d"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.906978 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d","Type":"ContainerStarted","Data":"15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.907016 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d","Type":"ContainerStarted","Data":"10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.907030 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d","Type":"ContainerStarted","Data":"0cbff9fc12174144cc719e60fbbe7b582434ca878c5ceab392476c3c38c86d30"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.909676 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d3d092de-ab7b-4cd0-a939-20afa3d0d999","Type":"ContainerStarted","Data":"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.909728 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d3d092de-ab7b-4cd0-a939-20afa3d0d999","Type":"ContainerStarted","Data":"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.909736 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-log" containerID="cri-o://6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284" gracePeriod=30 Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.909830 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-httpd" containerID="cri-o://fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190" gracePeriod=30 Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.909746 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d3d092de-ab7b-4cd0-a939-20afa3d0d999","Type":"ContainerStarted","Data":"9f116015301144a22decf38c5f61207b1014b7465c8b48ed099fa15ecb52f770"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.912128 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"b2f1f7fc-9f01-4487-aec2-c5a34d501890","Type":"ContainerStarted","Data":"ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.912156 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"b2f1f7fc-9f01-4487-aec2-c5a34d501890","Type":"ContainerStarted","Data":"8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.912169 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"b2f1f7fc-9f01-4487-aec2-c5a34d501890","Type":"ContainerStarted","Data":"c36809e94ba1fb63f1a9651d25d0683d9aac975ee1c68c8f7d58ad8271000c92"} Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.927871 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=1.92784545 podStartE2EDuration="1.92784545s" podCreationTimestamp="2025-11-27 07:30:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:39.923417387 +0000 UTC m=+1323.813008217" watchObservedRunningTime="2025-11-27 07:30:39.92784545 +0000 UTC m=+1323.817436260" Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.961767 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=2.961751533 podStartE2EDuration="2.961751533s" podCreationTimestamp="2025-11-27 07:30:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:39.956065903 +0000 UTC m=+1323.845656713" watchObservedRunningTime="2025-11-27 07:30:39.961751533 +0000 UTC m=+1323.851342343" Nov 27 07:30:39 crc kubenswrapper[4706]: I1127 07:30:39.982240 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.982200169 podStartE2EDuration="2.982200169s" podCreationTimestamp="2025-11-27 07:30:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:39.977712703 +0000 UTC m=+1323.867303513" watchObservedRunningTime="2025-11-27 07:30:39.982200169 +0000 UTC m=+1323.871790979" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.018973 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.018953686 podStartE2EDuration="3.018953686s" podCreationTimestamp="2025-11-27 07:30:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:40.00660734 +0000 UTC m=+1323.896198160" watchObservedRunningTime="2025-11-27 07:30:40.018953686 +0000 UTC m=+1323.908544506" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.389474 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.455819 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffsq5\" (UniqueName: \"kubernetes.io/projected/d3d092de-ab7b-4cd0-a939-20afa3d0d999-kube-api-access-ffsq5\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456279 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-sys\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456316 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456365 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-var-locks-brick\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456369 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-sys" (OuterVolumeSpecName: "sys") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456385 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-config-data\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456449 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456472 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-dev\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456501 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-nvme\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456535 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-logs\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456549 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-lib-modules\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456576 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-scripts\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456597 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-iscsi\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456618 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-run\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456640 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-httpd-run\") pod \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\" (UID: \"d3d092de-ab7b-4cd0-a939-20afa3d0d999\") " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.456883 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457242 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457275 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457370 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457425 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-run" (OuterVolumeSpecName: "run") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457451 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457475 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-logs" (OuterVolumeSpecName: "logs") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457480 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-dev" (OuterVolumeSpecName: "dev") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.457512 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.462810 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-scripts" (OuterVolumeSpecName: "scripts") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.463714 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.465175 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.476487 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3d092de-ab7b-4cd0-a939-20afa3d0d999-kube-api-access-ffsq5" (OuterVolumeSpecName: "kube-api-access-ffsq5") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "kube-api-access-ffsq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.519781 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-config-data" (OuterVolumeSpecName: "config-data") pod "d3d092de-ab7b-4cd0-a939-20afa3d0d999" (UID: "d3d092de-ab7b-4cd0-a939-20afa3d0d999"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558534 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558584 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558596 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558606 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558619 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558630 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558641 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3d092de-ab7b-4cd0-a939-20afa3d0d999-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558652 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffsq5\" (UniqueName: \"kubernetes.io/projected/d3d092de-ab7b-4cd0-a939-20afa3d0d999-kube-api-access-ffsq5\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558698 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558711 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558723 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d092de-ab7b-4cd0-a939-20afa3d0d999-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558742 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.558754 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d3d092de-ab7b-4cd0-a939-20afa3d0d999-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.573756 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.576038 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.660306 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.660564 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.924808 4706 generic.go:334] "Generic (PLEG): container finished" podID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerID="fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190" exitCode=143 Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.925125 4706 generic.go:334] "Generic (PLEG): container finished" podID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerID="6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284" exitCode=143 Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.925860 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.926279 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d3d092de-ab7b-4cd0-a939-20afa3d0d999","Type":"ContainerDied","Data":"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190"} Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.926306 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d3d092de-ab7b-4cd0-a939-20afa3d0d999","Type":"ContainerDied","Data":"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284"} Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.926317 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d3d092de-ab7b-4cd0-a939-20afa3d0d999","Type":"ContainerDied","Data":"9f116015301144a22decf38c5f61207b1014b7465c8b48ed099fa15ecb52f770"} Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.926333 4706 scope.go:117] "RemoveContainer" containerID="fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.952763 4706 scope.go:117] "RemoveContainer" containerID="6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.952851 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.958088 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.980411 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:40 crc kubenswrapper[4706]: E1127 07:30:40.980678 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-httpd" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.980690 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-httpd" Nov 27 07:30:40 crc kubenswrapper[4706]: E1127 07:30:40.980723 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-log" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.980732 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-log" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.980854 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-httpd" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.980880 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" containerName="glance-log" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.981546 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.984196 4706 scope.go:117] "RemoveContainer" containerID="fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190" Nov 27 07:30:40 crc kubenswrapper[4706]: E1127 07:30:40.985108 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190\": container with ID starting with fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190 not found: ID does not exist" containerID="fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.985150 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190"} err="failed to get container status \"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190\": rpc error: code = NotFound desc = could not find container \"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190\": container with ID starting with fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190 not found: ID does not exist" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.985172 4706 scope.go:117] "RemoveContainer" containerID="6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284" Nov 27 07:30:40 crc kubenswrapper[4706]: E1127 07:30:40.985483 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284\": container with ID starting with 6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284 not found: ID does not exist" containerID="6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.985506 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284"} err="failed to get container status \"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284\": rpc error: code = NotFound desc = could not find container \"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284\": container with ID starting with 6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284 not found: ID does not exist" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.985527 4706 scope.go:117] "RemoveContainer" containerID="fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.985787 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190"} err="failed to get container status \"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190\": rpc error: code = NotFound desc = could not find container \"fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190\": container with ID starting with fa7f3a1c00615cad4a797457e57523f948989ea3b15f21e16a3b7516b0ea1190 not found: ID does not exist" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.985808 4706 scope.go:117] "RemoveContainer" containerID="6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284" Nov 27 07:30:40 crc kubenswrapper[4706]: I1127 07:30:40.986014 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284"} err="failed to get container status \"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284\": rpc error: code = NotFound desc = could not find container \"6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284\": container with ID starting with 6243f7574bda841701cde07f7bff8e6ab17fef101cd7e3f7a8841fd83710d284 not found: ID does not exist" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.008112 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169498 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169580 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-dev\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169614 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169642 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169692 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169847 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdfmv\" (UniqueName: \"kubernetes.io/projected/f5e779bc-2eae-44aa-890f-bb9c36aac432-kube-api-access-rdfmv\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169893 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-sys\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169916 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-scripts\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.169985 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.170023 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-logs\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.170113 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-run\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.170150 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-config-data\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.170173 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.170334 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.272905 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273305 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273048 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273340 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-dev\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273385 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-dev\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273410 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273444 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273474 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273492 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdfmv\" (UniqueName: \"kubernetes.io/projected/f5e779bc-2eae-44aa-890f-bb9c36aac432-kube-api-access-rdfmv\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273520 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-sys\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273535 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-scripts\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273551 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273558 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273582 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-logs\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273609 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-run\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273626 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-config-data\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273646 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273710 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.273949 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.274037 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.274172 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.274269 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-logs\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.274301 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-run\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.274601 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-sys\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.282593 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-scripts\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.282902 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-config-data\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.299698 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdfmv\" (UniqueName: \"kubernetes.io/projected/f5e779bc-2eae-44aa-890f-bb9c36aac432-kube-api-access-rdfmv\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.301609 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.304749 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:41 crc kubenswrapper[4706]: I1127 07:30:41.608533 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:42 crc kubenswrapper[4706]: I1127 07:30:42.164628 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:30:42 crc kubenswrapper[4706]: I1127 07:30:42.793682 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3d092de-ab7b-4cd0-a939-20afa3d0d999" path="/var/lib/kubelet/pods/d3d092de-ab7b-4cd0-a939-20afa3d0d999/volumes" Nov 27 07:30:42 crc kubenswrapper[4706]: I1127 07:30:42.949529 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"f5e779bc-2eae-44aa-890f-bb9c36aac432","Type":"ContainerStarted","Data":"0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f"} Nov 27 07:30:42 crc kubenswrapper[4706]: I1127 07:30:42.949578 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"f5e779bc-2eae-44aa-890f-bb9c36aac432","Type":"ContainerStarted","Data":"3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa"} Nov 27 07:30:42 crc kubenswrapper[4706]: I1127 07:30:42.949594 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"f5e779bc-2eae-44aa-890f-bb9c36aac432","Type":"ContainerStarted","Data":"864a17f67d9aa37a0c338f0c37d077b42a677d31b73b209bb5053f2a67d064c9"} Nov 27 07:30:42 crc kubenswrapper[4706]: I1127 07:30:42.991405 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=2.991381715 podStartE2EDuration="2.991381715s" podCreationTimestamp="2025-11-27 07:30:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:42.982759627 +0000 UTC m=+1326.872350447" watchObservedRunningTime="2025-11-27 07:30:42.991381715 +0000 UTC m=+1326.880972535" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.385665 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.386430 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.426206 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.453324 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.471956 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.472896 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.530323 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.532386 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.720788 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.720855 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.748840 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:48 crc kubenswrapper[4706]: I1127 07:30:48.762802 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:49 crc kubenswrapper[4706]: I1127 07:30:49.006701 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:49 crc kubenswrapper[4706]: I1127 07:30:49.007527 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:49 crc kubenswrapper[4706]: I1127 07:30:49.007591 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:49 crc kubenswrapper[4706]: I1127 07:30:49.007637 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:49 crc kubenswrapper[4706]: I1127 07:30:49.007763 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:49 crc kubenswrapper[4706]: I1127 07:30:49.008644 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:50 crc kubenswrapper[4706]: I1127 07:30:50.814703 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:50 crc kubenswrapper[4706]: I1127 07:30:50.874707 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:50 crc kubenswrapper[4706]: I1127 07:30:50.884331 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:50 crc kubenswrapper[4706]: I1127 07:30:50.887595 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:30:50 crc kubenswrapper[4706]: I1127 07:30:50.895780 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:50 crc kubenswrapper[4706]: I1127 07:30:50.927775 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:50 crc kubenswrapper[4706]: I1127 07:30:50.940318 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:51 crc kubenswrapper[4706]: I1127 07:30:51.610362 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:51 crc kubenswrapper[4706]: I1127 07:30:51.610402 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:51 crc kubenswrapper[4706]: I1127 07:30:51.638927 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:51 crc kubenswrapper[4706]: I1127 07:30:51.651752 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:52 crc kubenswrapper[4706]: I1127 07:30:52.030767 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-log" containerID="cri-o://8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203" gracePeriod=30 Nov 27 07:30:52 crc kubenswrapper[4706]: I1127 07:30:52.030942 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-httpd" containerID="cri-o://ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f" gracePeriod=30 Nov 27 07:30:52 crc kubenswrapper[4706]: I1127 07:30:52.031107 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:52 crc kubenswrapper[4706]: I1127 07:30:52.031129 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:52 crc kubenswrapper[4706]: I1127 07:30:52.037748 4706 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.148:9292/healthcheck\": EOF" Nov 27 07:30:53 crc kubenswrapper[4706]: I1127 07:30:53.038249 4706 generic.go:334] "Generic (PLEG): container finished" podID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerID="8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203" exitCode=143 Nov 27 07:30:53 crc kubenswrapper[4706]: I1127 07:30:53.038501 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"b2f1f7fc-9f01-4487-aec2-c5a34d501890","Type":"ContainerDied","Data":"8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203"} Nov 27 07:30:53 crc kubenswrapper[4706]: I1127 07:30:53.858802 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:54 crc kubenswrapper[4706]: I1127 07:30:54.016766 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:30:54 crc kubenswrapper[4706]: I1127 07:30:54.084251 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:54 crc kubenswrapper[4706]: I1127 07:30:54.084572 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-log" containerID="cri-o://10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7" gracePeriod=30 Nov 27 07:30:54 crc kubenswrapper[4706]: I1127 07:30:54.084669 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-httpd" containerID="cri-o://15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd" gracePeriod=30 Nov 27 07:30:55 crc kubenswrapper[4706]: I1127 07:30:55.058591 4706 generic.go:334] "Generic (PLEG): container finished" podID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerID="10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7" exitCode=143 Nov 27 07:30:55 crc kubenswrapper[4706]: I1127 07:30:55.058640 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d","Type":"ContainerDied","Data":"10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7"} Nov 27 07:30:56 crc kubenswrapper[4706]: I1127 07:30:56.894904 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048444 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-httpd-run\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048540 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-nvme\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048614 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-iscsi\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048670 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-run\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048711 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjkmf\" (UniqueName: \"kubernetes.io/projected/b2f1f7fc-9f01-4487-aec2-c5a34d501890-kube-api-access-rjkmf\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048727 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048737 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048769 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-run" (OuterVolumeSpecName: "run") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048766 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-sys\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048823 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-var-locks-brick\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048816 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048865 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048846 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-sys" (OuterVolumeSpecName: "sys") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048857 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-config-data\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.048978 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-scripts\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049092 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049111 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049125 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-dev\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049151 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-lib-modules\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049174 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-logs\") pod \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\" (UID: \"b2f1f7fc-9f01-4487-aec2-c5a34d501890\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049712 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049725 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049734 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049744 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049751 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049759 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.049998 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-logs" (OuterVolumeSpecName: "logs") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.050024 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-dev" (OuterVolumeSpecName: "dev") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.050042 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.055310 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.055335 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2f1f7fc-9f01-4487-aec2-c5a34d501890-kube-api-access-rjkmf" (OuterVolumeSpecName: "kube-api-access-rjkmf") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "kube-api-access-rjkmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.067472 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.070372 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-scripts" (OuterVolumeSpecName: "scripts") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.083641 4706 generic.go:334] "Generic (PLEG): container finished" podID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerID="ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f" exitCode=0 Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.083719 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.083737 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"b2f1f7fc-9f01-4487-aec2-c5a34d501890","Type":"ContainerDied","Data":"ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f"} Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.084131 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"b2f1f7fc-9f01-4487-aec2-c5a34d501890","Type":"ContainerDied","Data":"c36809e94ba1fb63f1a9651d25d0683d9aac975ee1c68c8f7d58ad8271000c92"} Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.084166 4706 scope.go:117] "RemoveContainer" containerID="ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.142040 4706 scope.go:117] "RemoveContainer" containerID="8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.145599 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-config-data" (OuterVolumeSpecName: "config-data") pod "b2f1f7fc-9f01-4487-aec2-c5a34d501890" (UID: "b2f1f7fc-9f01-4487-aec2-c5a34d501890"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151050 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151079 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f1f7fc-9f01-4487-aec2-c5a34d501890-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151089 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjkmf\" (UniqueName: \"kubernetes.io/projected/b2f1f7fc-9f01-4487-aec2-c5a34d501890-kube-api-access-rjkmf\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151100 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151109 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f1f7fc-9f01-4487-aec2-c5a34d501890-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151134 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151144 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b2f1f7fc-9f01-4487-aec2-c5a34d501890-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.151156 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.186595 4706 scope.go:117] "RemoveContainer" containerID="ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f" Nov 27 07:30:57 crc kubenswrapper[4706]: E1127 07:30:57.187003 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f\": container with ID starting with ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f not found: ID does not exist" containerID="ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.187027 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f"} err="failed to get container status \"ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f\": rpc error: code = NotFound desc = could not find container \"ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f\": container with ID starting with ce318fdcc624f7e76a9a90d6232a5abb37e35daa26591dc99ff0aee5f1a83b5f not found: ID does not exist" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.187047 4706 scope.go:117] "RemoveContainer" containerID="8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203" Nov 27 07:30:57 crc kubenswrapper[4706]: E1127 07:30:57.187594 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203\": container with ID starting with 8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203 not found: ID does not exist" containerID="8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.187612 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203"} err="failed to get container status \"8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203\": rpc error: code = NotFound desc = could not find container \"8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203\": container with ID starting with 8202e3cc07f64d1868f027161b0f84cc7032a3bf29b77af75067bd5197917203 not found: ID does not exist" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.191641 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.203284 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.252019 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.252044 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.424175 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.435360 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.448155 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:57 crc kubenswrapper[4706]: E1127 07:30:57.448449 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-httpd" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.448466 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-httpd" Nov 27 07:30:57 crc kubenswrapper[4706]: E1127 07:30:57.448490 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-log" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.448498 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-log" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.448617 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-log" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.448633 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" containerName="glance-httpd" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.449299 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.464826 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.529825 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560734 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-dev\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560794 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-run\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560815 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-logs\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560836 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560857 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560878 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-config-data\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560902 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fzw8\" (UniqueName: \"kubernetes.io/projected/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-kube-api-access-8fzw8\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560922 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-sys\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560945 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560964 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.560995 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.561012 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.561033 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-scripts\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.561050 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662326 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnszx\" (UniqueName: \"kubernetes.io/projected/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-kube-api-access-jnszx\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662381 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-lib-modules\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662399 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-sys\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662413 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-iscsi\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662436 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-run\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662455 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-httpd-run\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662497 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-dev\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662534 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662554 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-logs\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662551 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-sys" (OuterVolumeSpecName: "sys") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662581 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-run" (OuterVolumeSpecName: "run") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662576 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662616 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-dev" (OuterVolumeSpecName: "dev") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662620 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662583 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-var-locks-brick\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662658 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662708 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662753 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-config-data\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662798 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-scripts\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662838 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-nvme\") pod \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\" (UID: \"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d\") " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662925 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.662969 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-logs" (OuterVolumeSpecName: "logs") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663058 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fzw8\" (UniqueName: \"kubernetes.io/projected/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-kube-api-access-8fzw8\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663071 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663113 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-sys\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663166 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663177 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-sys\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663241 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663310 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663333 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663365 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-scripts\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663382 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663424 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-dev\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663433 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663489 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-run\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663504 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663513 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-logs\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663547 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663579 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663605 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-config-data\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663651 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663662 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663671 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663680 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663688 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663697 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663706 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663714 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.663722 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664241 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-dev\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664277 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664294 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664568 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664662 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-logs\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664738 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664748 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.664837 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-run\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.666308 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.666326 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.666348 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-scripts" (OuterVolumeSpecName: "scripts") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.668493 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-kube-api-access-jnszx" (OuterVolumeSpecName: "kube-api-access-jnszx") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "kube-api-access-jnszx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.670466 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-scripts\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.675861 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-config-data\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.685343 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.691367 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fzw8\" (UniqueName: \"kubernetes.io/projected/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-kube-api-access-8fzw8\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.707060 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-0\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.712311 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-config-data" (OuterVolumeSpecName: "config-data") pod "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" (UID: "a0f52bf2-bba2-46d6-b97a-3c36f800ae4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.764972 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.765012 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.765024 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.765034 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.765067 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnszx\" (UniqueName: \"kubernetes.io/projected/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d-kube-api-access-jnszx\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.770679 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.777545 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.781127 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.866667 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:57 crc kubenswrapper[4706]: I1127 07:30:57.866951 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.094861 4706 generic.go:334] "Generic (PLEG): container finished" podID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerID="15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd" exitCode=0 Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.094927 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.094983 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d","Type":"ContainerDied","Data":"15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd"} Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.095057 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a0f52bf2-bba2-46d6-b97a-3c36f800ae4d","Type":"ContainerDied","Data":"0cbff9fc12174144cc719e60fbbe7b582434ca878c5ceab392476c3c38c86d30"} Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.095361 4706 scope.go:117] "RemoveContainer" containerID="15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.121574 4706 scope.go:117] "RemoveContainer" containerID="10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.130565 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.148529 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.163672 4706 scope.go:117] "RemoveContainer" containerID="15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd" Nov 27 07:30:58 crc kubenswrapper[4706]: E1127 07:30:58.165196 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd\": container with ID starting with 15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd not found: ID does not exist" containerID="15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.165310 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd"} err="failed to get container status \"15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd\": rpc error: code = NotFound desc = could not find container \"15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd\": container with ID starting with 15f5114f99a5e328559a06cb07663753c38ef940c81f929de200eb409cf3dbfd not found: ID does not exist" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.165356 4706 scope.go:117] "RemoveContainer" containerID="10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7" Nov 27 07:30:58 crc kubenswrapper[4706]: E1127 07:30:58.165874 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7\": container with ID starting with 10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7 not found: ID does not exist" containerID="10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.165929 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7"} err="failed to get container status \"10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7\": rpc error: code = NotFound desc = could not find container \"10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7\": container with ID starting with 10f3215a22ac22d501206b83d7cc5a0a2cc551d89c06c1eea29925de842b55d7 not found: ID does not exist" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.170966 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:58 crc kubenswrapper[4706]: E1127 07:30:58.171430 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-httpd" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.171457 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-httpd" Nov 27 07:30:58 crc kubenswrapper[4706]: E1127 07:30:58.171488 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-log" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.171503 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-log" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.171723 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-httpd" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.171757 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" containerName="glance-log" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.173044 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.180202 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.230047 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:30:58 crc kubenswrapper[4706]: W1127 07:30:58.233427 4706 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb4b7d2b_970b_410d_b11c_e11ac0b8a4bb.slice/crio-d144942dd56504b45cf3c62eb40f3f4e0a79d96bfb18680b8495e9bbf3b1b6f4 WatchSource:0}: Error finding container d144942dd56504b45cf3c62eb40f3f4e0a79d96bfb18680b8495e9bbf3b1b6f4: Status 404 returned error can't find the container with id d144942dd56504b45cf3c62eb40f3f4e0a79d96bfb18680b8495e9bbf3b1b6f4 Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.272934 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.272969 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.272999 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273016 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8tp9\" (UniqueName: \"kubernetes.io/projected/5b99e235-cb57-43e7-a63e-2f177dc7549d-kube-api-access-q8tp9\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273042 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273059 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273078 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273163 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273214 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273286 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-sys\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273326 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-run\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273417 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-dev\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273483 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.273505 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-logs\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375063 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375332 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-logs\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375380 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375398 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375424 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375438 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8tp9\" (UniqueName: \"kubernetes.io/projected/5b99e235-cb57-43e7-a63e-2f177dc7549d-kube-api-access-q8tp9\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375462 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375479 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375496 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375513 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375530 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375547 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-sys\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375567 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-run\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375592 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-dev\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375607 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-logs\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375657 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375664 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-dev\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.375829 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.376092 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.376139 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.376163 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-run\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.376185 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.376183 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-sys\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.376412 4706 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.376452 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.379198 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.380401 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.389916 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8tp9\" (UniqueName: \"kubernetes.io/projected/5b99e235-cb57-43e7-a63e-2f177dc7549d-kube-api-access-q8tp9\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.395186 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.395491 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.489302 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.785892 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f52bf2-bba2-46d6-b97a-3c36f800ae4d" path="/var/lib/kubelet/pods/a0f52bf2-bba2-46d6-b97a-3c36f800ae4d/volumes" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.787160 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2f1f7fc-9f01-4487-aec2-c5a34d501890" path="/var/lib/kubelet/pods/b2f1f7fc-9f01-4487-aec2-c5a34d501890/volumes" Nov 27 07:30:58 crc kubenswrapper[4706]: I1127 07:30:58.951112 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:30:59 crc kubenswrapper[4706]: I1127 07:30:59.109327 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5b99e235-cb57-43e7-a63e-2f177dc7549d","Type":"ContainerStarted","Data":"f5d1e20e46923051afa203e6e8941a4f3f92f6f51943bbbac37039bfc408469a"} Nov 27 07:30:59 crc kubenswrapper[4706]: I1127 07:30:59.112231 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb","Type":"ContainerStarted","Data":"d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d"} Nov 27 07:30:59 crc kubenswrapper[4706]: I1127 07:30:59.112279 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb","Type":"ContainerStarted","Data":"684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e"} Nov 27 07:30:59 crc kubenswrapper[4706]: I1127 07:30:59.112292 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb","Type":"ContainerStarted","Data":"d144942dd56504b45cf3c62eb40f3f4e0a79d96bfb18680b8495e9bbf3b1b6f4"} Nov 27 07:31:00 crc kubenswrapper[4706]: I1127 07:31:00.131754 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5b99e235-cb57-43e7-a63e-2f177dc7549d","Type":"ContainerStarted","Data":"3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c"} Nov 27 07:31:00 crc kubenswrapper[4706]: I1127 07:31:00.132645 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5b99e235-cb57-43e7-a63e-2f177dc7549d","Type":"ContainerStarted","Data":"48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68"} Nov 27 07:31:00 crc kubenswrapper[4706]: I1127 07:31:00.166860 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=3.166833629 podStartE2EDuration="3.166833629s" podCreationTimestamp="2025-11-27 07:30:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:30:59.138606783 +0000 UTC m=+1343.028197603" watchObservedRunningTime="2025-11-27 07:31:00.166833629 +0000 UTC m=+1344.056424479" Nov 27 07:31:00 crc kubenswrapper[4706]: I1127 07:31:00.167387 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.167379139 podStartE2EDuration="2.167379139s" podCreationTimestamp="2025-11-27 07:30:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:31:00.159386544 +0000 UTC m=+1344.048977394" watchObservedRunningTime="2025-11-27 07:31:00.167379139 +0000 UTC m=+1344.056969989" Nov 27 07:31:07 crc kubenswrapper[4706]: I1127 07:31:07.771682 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:07 crc kubenswrapper[4706]: I1127 07:31:07.772367 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:07 crc kubenswrapper[4706]: I1127 07:31:07.820802 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:07 crc kubenswrapper[4706]: I1127 07:31:07.847094 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:08 crc kubenswrapper[4706]: I1127 07:31:08.204506 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:08 crc kubenswrapper[4706]: I1127 07:31:08.204574 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:08 crc kubenswrapper[4706]: I1127 07:31:08.490519 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:08 crc kubenswrapper[4706]: I1127 07:31:08.490565 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:08 crc kubenswrapper[4706]: I1127 07:31:08.530641 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:08 crc kubenswrapper[4706]: I1127 07:31:08.560900 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:09 crc kubenswrapper[4706]: I1127 07:31:09.213269 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:09 crc kubenswrapper[4706]: I1127 07:31:09.213313 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:10 crc kubenswrapper[4706]: I1127 07:31:10.044602 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:10 crc kubenswrapper[4706]: I1127 07:31:10.090565 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:10 crc kubenswrapper[4706]: I1127 07:31:10.985523 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:11 crc kubenswrapper[4706]: I1127 07:31:11.098547 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:37 crc kubenswrapper[4706]: I1127 07:31:37.952641 4706 scope.go:117] "RemoveContainer" containerID="fa800bf58c6ed5a8d809f8f879a27d936f6231c341043933b237d54fc820e5ce" Nov 27 07:31:37 crc kubenswrapper[4706]: I1127 07:31:37.973709 4706 scope.go:117] "RemoveContainer" containerID="c7f09509e6118431d8aa40dd5faf8bb79c5b34059e6435631f0f60e1c386fc32" Nov 27 07:31:38 crc kubenswrapper[4706]: I1127 07:31:38.035332 4706 scope.go:117] "RemoveContainer" containerID="c81e1bc161bc027db217be0c897cb083f59d3e9f287e621bd6d1eb3174b57ac3" Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.362104 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.363009 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-log" containerID="cri-o://dd37542636329515e16d92d078921902e2eb76643303dd071ecb13da84e7438d" gracePeriod=30 Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.363112 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-httpd" containerID="cri-o://8967b07927196e003fbc8d9d9c7b1573d12469c74464c3207c0b7aa091a96cbf" gracePeriod=30 Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.515943 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.516443 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-log" containerID="cri-o://3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa" gracePeriod=30 Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.516594 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-httpd" containerID="cri-o://0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f" gracePeriod=30 Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.550370 4706 generic.go:334] "Generic (PLEG): container finished" podID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerID="dd37542636329515e16d92d078921902e2eb76643303dd071ecb13da84e7438d" exitCode=143 Nov 27 07:31:48 crc kubenswrapper[4706]: I1127 07:31:48.550417 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"5dfa1095-7cae-4bdd-9615-d9a447134b6f","Type":"ContainerDied","Data":"dd37542636329515e16d92d078921902e2eb76643303dd071ecb13da84e7438d"} Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.564519 4706 generic.go:334] "Generic (PLEG): container finished" podID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerID="3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa" exitCode=143 Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.564562 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"f5e779bc-2eae-44aa-890f-bb9c36aac432","Type":"ContainerDied","Data":"3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa"} Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.739341 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cskxq"] Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.745814 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cskxq"] Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.793685 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance6f10-account-delete-4wzj4"] Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.795088 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.801617 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance6f10-account-delete-4wzj4"] Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.840819 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.841050 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-log" containerID="cri-o://684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e" gracePeriod=30 Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.841148 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-httpd" containerID="cri-o://d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d" gracePeriod=30 Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.903204 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.903632 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-log" containerID="cri-o://48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68" gracePeriod=30 Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.903856 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-httpd" containerID="cri-o://3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c" gracePeriod=30 Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.943736 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e2b34ae-f2f3-4571-bc32-3f002a955862-operator-scripts\") pod \"glance6f10-account-delete-4wzj4\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:49 crc kubenswrapper[4706]: I1127 07:31:49.943831 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mqt9\" (UniqueName: \"kubernetes.io/projected/7e2b34ae-f2f3-4571-bc32-3f002a955862-kube-api-access-8mqt9\") pod \"glance6f10-account-delete-4wzj4\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.044869 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e2b34ae-f2f3-4571-bc32-3f002a955862-operator-scripts\") pod \"glance6f10-account-delete-4wzj4\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.045331 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mqt9\" (UniqueName: \"kubernetes.io/projected/7e2b34ae-f2f3-4571-bc32-3f002a955862-kube-api-access-8mqt9\") pod \"glance6f10-account-delete-4wzj4\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.045901 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e2b34ae-f2f3-4571-bc32-3f002a955862-operator-scripts\") pod \"glance6f10-account-delete-4wzj4\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.065202 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mqt9\" (UniqueName: \"kubernetes.io/projected/7e2b34ae-f2f3-4571-bc32-3f002a955862-kube-api-access-8mqt9\") pod \"glance6f10-account-delete-4wzj4\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.116774 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.386340 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance6f10-account-delete-4wzj4"] Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.573482 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" event={"ID":"7e2b34ae-f2f3-4571-bc32-3f002a955862","Type":"ContainerStarted","Data":"31d6c1ade58f373e3a672e680187bff8f7098ad5d1c6f2cc60431e79fb4927ca"} Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.574548 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" event={"ID":"7e2b34ae-f2f3-4571-bc32-3f002a955862","Type":"ContainerStarted","Data":"3e85dc712772f64fdd9e3c5c75f838e2e14a9ac50eca9b4af6e479f597e3ae0a"} Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.576630 4706 generic.go:334] "Generic (PLEG): container finished" podID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerID="48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68" exitCode=143 Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.576714 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5b99e235-cb57-43e7-a63e-2f177dc7549d","Type":"ContainerDied","Data":"48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68"} Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.579376 4706 generic.go:334] "Generic (PLEG): container finished" podID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerID="684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e" exitCode=143 Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.579412 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb","Type":"ContainerDied","Data":"684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e"} Nov 27 07:31:50 crc kubenswrapper[4706]: I1127 07:31:50.786995 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8b675cb-a460-4c74-81bf-80223ae396f2" path="/var/lib/kubelet/pods/d8b675cb-a460-4c74-81bf-80223ae396f2/volumes" Nov 27 07:31:51 crc kubenswrapper[4706]: I1127 07:31:51.598522 4706 generic.go:334] "Generic (PLEG): container finished" podID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerID="8967b07927196e003fbc8d9d9c7b1573d12469c74464c3207c0b7aa091a96cbf" exitCode=0 Nov 27 07:31:51 crc kubenswrapper[4706]: I1127 07:31:51.598595 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"5dfa1095-7cae-4bdd-9615-d9a447134b6f","Type":"ContainerDied","Data":"8967b07927196e003fbc8d9d9c7b1573d12469c74464c3207c0b7aa091a96cbf"} Nov 27 07:31:51 crc kubenswrapper[4706]: I1127 07:31:51.604877 4706 generic.go:334] "Generic (PLEG): container finished" podID="7e2b34ae-f2f3-4571-bc32-3f002a955862" containerID="31d6c1ade58f373e3a672e680187bff8f7098ad5d1c6f2cc60431e79fb4927ca" exitCode=0 Nov 27 07:31:51 crc kubenswrapper[4706]: I1127 07:31:51.604922 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" event={"ID":"7e2b34ae-f2f3-4571-bc32-3f002a955862","Type":"ContainerDied","Data":"31d6c1ade58f373e3a672e680187bff8f7098ad5d1c6f2cc60431e79fb4927ca"} Nov 27 07:31:51 crc kubenswrapper[4706]: I1127 07:31:51.925443 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.040796 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080255 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-run\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080326 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-nvme\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080343 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-run" (OuterVolumeSpecName: "run") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080353 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-config-data\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080421 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-dev\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080427 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080474 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-dev" (OuterVolumeSpecName: "dev") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080489 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-logs\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080526 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080572 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-scripts\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080598 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7mrc\" (UniqueName: \"kubernetes.io/projected/5dfa1095-7cae-4bdd-9615-d9a447134b6f-kube-api-access-s7mrc\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080624 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-httpd-run\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080653 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-sys\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080697 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-var-locks-brick\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080734 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-lib-modules\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080738 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-logs" (OuterVolumeSpecName: "logs") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080760 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-iscsi\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080789 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\" (UID: \"5dfa1095-7cae-4bdd-9615-d9a447134b6f\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.080879 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081001 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081029 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-sys" (OuterVolumeSpecName: "sys") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081072 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081118 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081255 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081280 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081290 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081299 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5dfa1095-7cae-4bdd-9615-d9a447134b6f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081310 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081320 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081331 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081340 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.081349 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dfa1095-7cae-4bdd-9615-d9a447134b6f-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.087528 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-scripts" (OuterVolumeSpecName: "scripts") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.087540 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance-cache") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.087556 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dfa1095-7cae-4bdd-9615-d9a447134b6f-kube-api-access-s7mrc" (OuterVolumeSpecName: "kube-api-access-s7mrc") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "kube-api-access-s7mrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.090405 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.124394 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-config-data" (OuterVolumeSpecName: "config-data") pod "5dfa1095-7cae-4bdd-9615-d9a447134b6f" (UID: "5dfa1095-7cae-4bdd-9615-d9a447134b6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182341 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-sys\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182424 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182499 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-scripts\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182506 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-sys" (OuterVolumeSpecName: "sys") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182525 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182606 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-var-locks-brick\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182654 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-run\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182717 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-iscsi\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182774 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdfmv\" (UniqueName: \"kubernetes.io/projected/f5e779bc-2eae-44aa-890f-bb9c36aac432-kube-api-access-rdfmv\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182806 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-httpd-run\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182855 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-dev\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182860 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-run" (OuterVolumeSpecName: "run") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182886 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-nvme\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182924 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182924 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-lib-modules\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182953 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.182984 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183011 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-config-data\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183035 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-logs\") pod \"f5e779bc-2eae-44aa-890f-bb9c36aac432\" (UID: \"f5e779bc-2eae-44aa-890f-bb9c36aac432\") " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183546 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-dev" (OuterVolumeSpecName: "dev") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183608 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183660 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183672 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183682 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183692 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7mrc\" (UniqueName: \"kubernetes.io/projected/5dfa1095-7cae-4bdd-9615-d9a447134b6f-kube-api-access-s7mrc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183705 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183715 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183734 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183747 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dfa1095-7cae-4bdd-9615-d9a447134b6f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183782 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183882 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.183946 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.184350 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-logs" (OuterVolumeSpecName: "logs") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.185953 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-scripts" (OuterVolumeSpecName: "scripts") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.186052 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.186690 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.187077 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e779bc-2eae-44aa-890f-bb9c36aac432-kube-api-access-rdfmv" (OuterVolumeSpecName: "kube-api-access-rdfmv") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "kube-api-access-rdfmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.199422 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.209673 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.233394 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-config-data" (OuterVolumeSpecName: "config-data") pod "f5e779bc-2eae-44aa-890f-bb9c36aac432" (UID: "f5e779bc-2eae-44aa-890f-bb9c36aac432"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285684 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285720 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285732 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f5e779bc-2eae-44aa-890f-bb9c36aac432-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285745 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285792 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285803 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e779bc-2eae-44aa-890f-bb9c36aac432-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285840 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285920 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5e779bc-2eae-44aa-890f-bb9c36aac432-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285957 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285970 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.285984 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdfmv\" (UniqueName: \"kubernetes.io/projected/f5e779bc-2eae-44aa-890f-bb9c36aac432-kube-api-access-rdfmv\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.303290 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.303336 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.387641 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.387672 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.619042 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"5dfa1095-7cae-4bdd-9615-d9a447134b6f","Type":"ContainerDied","Data":"68da73b505ac9edbc564772d42ef87c2e02a2e18a78e430864b988e5db158080"} Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.619098 4706 scope.go:117] "RemoveContainer" containerID="8967b07927196e003fbc8d9d9c7b1573d12469c74464c3207c0b7aa091a96cbf" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.619262 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.623851 4706 generic.go:334] "Generic (PLEG): container finished" podID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerID="0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f" exitCode=0 Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.623907 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.623938 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"f5e779bc-2eae-44aa-890f-bb9c36aac432","Type":"ContainerDied","Data":"0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f"} Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.623986 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"f5e779bc-2eae-44aa-890f-bb9c36aac432","Type":"ContainerDied","Data":"864a17f67d9aa37a0c338f0c37d077b42a677d31b73b209bb5053f2a67d064c9"} Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.646038 4706 scope.go:117] "RemoveContainer" containerID="dd37542636329515e16d92d078921902e2eb76643303dd071ecb13da84e7438d" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.658099 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.675027 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.679992 4706 scope.go:117] "RemoveContainer" containerID="0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.682302 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.688521 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.705817 4706 scope.go:117] "RemoveContainer" containerID="3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.721802 4706 scope.go:117] "RemoveContainer" containerID="0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f" Nov 27 07:31:52 crc kubenswrapper[4706]: E1127 07:31:52.722196 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f\": container with ID starting with 0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f not found: ID does not exist" containerID="0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.722255 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f"} err="failed to get container status \"0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f\": rpc error: code = NotFound desc = could not find container \"0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f\": container with ID starting with 0857b18356d1ae7d747cde1c40c1dfa0e0d7f188993c883f1b0f6688e35dc01f not found: ID does not exist" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.722282 4706 scope.go:117] "RemoveContainer" containerID="3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa" Nov 27 07:31:52 crc kubenswrapper[4706]: E1127 07:31:52.722627 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa\": container with ID starting with 3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa not found: ID does not exist" containerID="3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.722658 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa"} err="failed to get container status \"3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa\": rpc error: code = NotFound desc = could not find container \"3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa\": container with ID starting with 3670dc773e192d88e432ae0ea9f8efee5cc94c3f4f0a8a4109be4dfd8eaf81aa not found: ID does not exist" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.787759 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" path="/var/lib/kubelet/pods/5dfa1095-7cae-4bdd-9615-d9a447134b6f/volumes" Nov 27 07:31:52 crc kubenswrapper[4706]: I1127 07:31:52.788665 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" path="/var/lib/kubelet/pods/f5e779bc-2eae-44aa-890f-bb9c36aac432/volumes" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.042871 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.197686 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e2b34ae-f2f3-4571-bc32-3f002a955862-operator-scripts\") pod \"7e2b34ae-f2f3-4571-bc32-3f002a955862\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.197725 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mqt9\" (UniqueName: \"kubernetes.io/projected/7e2b34ae-f2f3-4571-bc32-3f002a955862-kube-api-access-8mqt9\") pod \"7e2b34ae-f2f3-4571-bc32-3f002a955862\" (UID: \"7e2b34ae-f2f3-4571-bc32-3f002a955862\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.198425 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e2b34ae-f2f3-4571-bc32-3f002a955862-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e2b34ae-f2f3-4571-bc32-3f002a955862" (UID: "7e2b34ae-f2f3-4571-bc32-3f002a955862"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.201648 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2b34ae-f2f3-4571-bc32-3f002a955862-kube-api-access-8mqt9" (OuterVolumeSpecName: "kube-api-access-8mqt9") pod "7e2b34ae-f2f3-4571-bc32-3f002a955862" (UID: "7e2b34ae-f2f3-4571-bc32-3f002a955862"). InnerVolumeSpecName "kube-api-access-8mqt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.299502 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e2b34ae-f2f3-4571-bc32-3f002a955862-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.299530 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mqt9\" (UniqueName: \"kubernetes.io/projected/7e2b34ae-f2f3-4571-bc32-3f002a955862-kube-api-access-8mqt9\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.302783 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.363545 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400038 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-logs\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400081 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-iscsi\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400106 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-sys\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400174 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400243 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-var-locks-brick\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400265 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-nvme\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400279 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-sys" (OuterVolumeSpecName: "sys") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400290 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-scripts\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400308 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400321 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400330 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400344 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-config-data\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400364 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400386 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-httpd-run\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400428 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-run\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400459 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-dev\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400474 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-logs" (OuterVolumeSpecName: "logs") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400484 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fzw8\" (UniqueName: \"kubernetes.io/projected/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-kube-api-access-8fzw8\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400510 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-lib-modules\") pod \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\" (UID: \"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400695 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-run" (OuterVolumeSpecName: "run") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400745 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-dev" (OuterVolumeSpecName: "dev") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400857 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400869 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400960 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400972 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400981 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400993 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.401005 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.400980 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.401807 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.403844 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.405296 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-scripts" (OuterVolumeSpecName: "scripts") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.407379 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-kube-api-access-8fzw8" (OuterVolumeSpecName: "kube-api-access-8fzw8") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "kube-api-access-8fzw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.408334 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.445781 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-config-data" (OuterVolumeSpecName: "config-data") pod "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" (UID: "cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501681 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-scripts\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501728 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-sys\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501774 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-nvme\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501790 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-dev\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501815 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-logs\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501832 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-var-locks-brick\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501868 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8tp9\" (UniqueName: \"kubernetes.io/projected/5b99e235-cb57-43e7-a63e-2f177dc7549d-kube-api-access-q8tp9\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501866 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501909 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501879 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-dev" (OuterVolumeSpecName: "dev") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501874 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-sys" (OuterVolumeSpecName: "sys") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501890 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-httpd-run\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.501992 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-lib-modules\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502052 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-iscsi\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502090 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-run\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502157 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-config-data\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502177 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502232 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502241 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"5b99e235-cb57-43e7-a63e-2f177dc7549d\" (UID: \"5b99e235-cb57-43e7-a63e-2f177dc7549d\") " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502273 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-run" (OuterVolumeSpecName: "run") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502294 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502312 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502342 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-logs" (OuterVolumeSpecName: "logs") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502780 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fzw8\" (UniqueName: \"kubernetes.io/projected/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-kube-api-access-8fzw8\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502800 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502811 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502821 4706 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502831 4706 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502841 4706 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502851 4706 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-sys\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502861 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502872 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502896 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502915 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502926 4706 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502937 4706 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502950 4706 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-dev\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502963 4706 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b99e235-cb57-43e7-a63e-2f177dc7549d-logs\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.502973 4706 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5b99e235-cb57-43e7-a63e-2f177dc7549d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.505412 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.506268 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b99e235-cb57-43e7-a63e-2f177dc7549d-kube-api-access-q8tp9" (OuterVolumeSpecName: "kube-api-access-q8tp9") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "kube-api-access-q8tp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.507621 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.508738 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-scripts" (OuterVolumeSpecName: "scripts") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.517490 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.517608 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.535646 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-config-data" (OuterVolumeSpecName: "config-data") pod "5b99e235-cb57-43e7-a63e-2f177dc7549d" (UID: "5b99e235-cb57-43e7-a63e-2f177dc7549d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.604427 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.604755 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.604765 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8tp9\" (UniqueName: \"kubernetes.io/projected/5b99e235-cb57-43e7-a63e-2f177dc7549d-kube-api-access-q8tp9\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.604776 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.604804 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.604817 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.604825 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99e235-cb57-43e7-a63e-2f177dc7549d-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.618138 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.618396 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.647432 4706 generic.go:334] "Generic (PLEG): container finished" podID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerID="3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c" exitCode=0 Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.647584 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.647737 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5b99e235-cb57-43e7-a63e-2f177dc7549d","Type":"ContainerDied","Data":"3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c"} Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.647804 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5b99e235-cb57-43e7-a63e-2f177dc7549d","Type":"ContainerDied","Data":"f5d1e20e46923051afa203e6e8941a4f3f92f6f51943bbbac37039bfc408469a"} Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.647823 4706 scope.go:117] "RemoveContainer" containerID="3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.651214 4706 generic.go:334] "Generic (PLEG): container finished" podID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerID="d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d" exitCode=0 Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.651339 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb","Type":"ContainerDied","Data":"d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d"} Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.651341 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.651623 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb","Type":"ContainerDied","Data":"d144942dd56504b45cf3c62eb40f3f4e0a79d96bfb18680b8495e9bbf3b1b6f4"} Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.655280 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" event={"ID":"7e2b34ae-f2f3-4571-bc32-3f002a955862","Type":"ContainerDied","Data":"3e85dc712772f64fdd9e3c5c75f838e2e14a9ac50eca9b4af6e479f597e3ae0a"} Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.655341 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e85dc712772f64fdd9e3c5c75f838e2e14a9ac50eca9b4af6e479f597e3ae0a" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.655405 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6f10-account-delete-4wzj4" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.668526 4706 scope.go:117] "RemoveContainer" containerID="48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.696193 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.698654 4706 scope.go:117] "RemoveContainer" containerID="3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c" Nov 27 07:31:53 crc kubenswrapper[4706]: E1127 07:31:53.699678 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c\": container with ID starting with 3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c not found: ID does not exist" containerID="3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.699878 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c"} err="failed to get container status \"3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c\": rpc error: code = NotFound desc = could not find container \"3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c\": container with ID starting with 3d9f167bd1884e34f0aaafe27731c1e9416ce4e1f968cd08e166f64b70615d7c not found: ID does not exist" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.699913 4706 scope.go:117] "RemoveContainer" containerID="48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68" Nov 27 07:31:53 crc kubenswrapper[4706]: E1127 07:31:53.700322 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68\": container with ID starting with 48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68 not found: ID does not exist" containerID="48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.700349 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68"} err="failed to get container status \"48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68\": rpc error: code = NotFound desc = could not find container \"48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68\": container with ID starting with 48b8d76c8c05cf1a967173ecb414c45d04b54d1398e7a31415397967cc837c68 not found: ID does not exist" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.700365 4706 scope.go:117] "RemoveContainer" containerID="d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.704690 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.706106 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.706138 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.708136 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.713641 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.721000 4706 scope.go:117] "RemoveContainer" containerID="684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.734730 4706 scope.go:117] "RemoveContainer" containerID="d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d" Nov 27 07:31:53 crc kubenswrapper[4706]: E1127 07:31:53.735239 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d\": container with ID starting with d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d not found: ID does not exist" containerID="d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.735281 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d"} err="failed to get container status \"d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d\": rpc error: code = NotFound desc = could not find container \"d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d\": container with ID starting with d8f8f18cf6f2b5eb7b9222458eaccde0890d6e8adca499a7b717dd2360ecb01d not found: ID does not exist" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.735309 4706 scope.go:117] "RemoveContainer" containerID="684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e" Nov 27 07:31:53 crc kubenswrapper[4706]: E1127 07:31:53.735636 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e\": container with ID starting with 684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e not found: ID does not exist" containerID="684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e" Nov 27 07:31:53 crc kubenswrapper[4706]: I1127 07:31:53.735697 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e"} err="failed to get container status \"684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e\": rpc error: code = NotFound desc = could not find container \"684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e\": container with ID starting with 684e6dadbf4c79db7296a491e4fc9cd2f514818849ce2b2d8a5bc1069872e77e not found: ID does not exist" Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.794959 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" path="/var/lib/kubelet/pods/5b99e235-cb57-43e7-a63e-2f177dc7549d/volumes" Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.796321 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" path="/var/lib/kubelet/pods/cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb/volumes" Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.824111 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-st6x5"] Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.829936 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-st6x5"] Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.841277 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance6f10-account-delete-4wzj4"] Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.847352 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-6f10-account-create-update-zmjz6"] Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.852198 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-6f10-account-create-update-zmjz6"] Nov 27 07:31:54 crc kubenswrapper[4706]: I1127 07:31:54.856846 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance6f10-account-delete-4wzj4"] Nov 27 07:31:56 crc kubenswrapper[4706]: I1127 07:31:56.785561 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f704bc5-b657-4454-b835-18c278b0bb24" path="/var/lib/kubelet/pods/5f704bc5-b657-4454-b835-18c278b0bb24/volumes" Nov 27 07:31:56 crc kubenswrapper[4706]: I1127 07:31:56.786656 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e2b34ae-f2f3-4571-bc32-3f002a955862" path="/var/lib/kubelet/pods/7e2b34ae-f2f3-4571-bc32-3f002a955862/volumes" Nov 27 07:31:56 crc kubenswrapper[4706]: I1127 07:31:56.787296 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96793f14-42e4-4a6f-8e18-cefa7e3f4bb6" path="/var/lib/kubelet/pods/96793f14-42e4-4a6f-8e18-cefa7e3f4bb6/volumes" Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.062793 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-2zfbf"] Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.072137 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-2zfbf"] Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087037 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087599 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-server" containerID="cri-o://be4e0386cd97f8aa803f6981bdc1c529197af63b34156666d8f8d7263d0d4f3d" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087686 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-server" containerID="cri-o://600192eff2ee82990dc4e3bfc25d3f3cdde9d6e9447251423860a8d59ffdffb8" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087733 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-updater" containerID="cri-o://e2ef352d28dc64cdc17f622c2a5ce4825d891748d60228a87b0234886b2fbd46" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087793 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-auditor" containerID="cri-o://de82cace4517a41df59b73d9a505b6d9bc00d26b88a1425eae7b033bb480c311" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087861 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-replicator" containerID="cri-o://de0a318ea00fadfa414f9b9cb4effedf4f8cbc78ffcf755bb0af2d1ee29bee87" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087918 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-server" containerID="cri-o://419c0ad19d57e0d824fc3545acc3c8fa4dc2b5da26669ee7c69100b12f6cf8ad" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087958 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-reaper" containerID="cri-o://02d59e38675cdde34dc486df647b1eff69092a995a977d61d28b1c5da1ad6b1e" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.087996 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-auditor" containerID="cri-o://b99dfd374a5ce1c13a934b2f873d212ff72dfe71b28341e2c047ff8d80e0c669" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.088039 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-replicator" containerID="cri-o://0854a4d87b476088ca363dd19e5424c00e3887a3cd94f9bb2324db18aa9630d0" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.088348 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-expirer" containerID="cri-o://13bb352c512127bb19cb77c2407f696591efaeb94de88485c71ef25ef7efd32b" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.088407 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="swift-recon-cron" containerID="cri-o://87334b52ad9d36330c1c6887ab6aefaab2f4a4d21ec015aff098d9b7df9b66e5" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.088448 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="rsync" containerID="cri-o://695ff04f9aad2b63580222fb53407fe88ab246ea589b941267e537a78643ddbd" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.088888 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-auditor" containerID="cri-o://b461410d2b38540e3971bd0c38fb4b2285b2b4bea2ad91957750db63eb8c977f" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.088935 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-updater" containerID="cri-o://2078680d2bb75db1546bb97e0d090d85113a906cd7a2b21900140fb580300a81" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.088985 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-replicator" containerID="cri-o://92f449c478541d4fc040be370d8fbfc4918ffb370dcd3a1fef9f2aeed113524a" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.123595 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs"] Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.124192 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-httpd" containerID="cri-o://9f23d04861cc7026fe4bf45c52c3a91306ee4fea16da518473dfc8059c16ef9c" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.124289 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-server" containerID="cri-o://c7619136346823904b27331aeb39977b3b1020d65267db3982003f2469a6f40c" gracePeriod=30 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750314 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="695ff04f9aad2b63580222fb53407fe88ab246ea589b941267e537a78643ddbd" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750675 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="13bb352c512127bb19cb77c2407f696591efaeb94de88485c71ef25ef7efd32b" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750686 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="2078680d2bb75db1546bb97e0d090d85113a906cd7a2b21900140fb580300a81" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750693 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="b461410d2b38540e3971bd0c38fb4b2285b2b4bea2ad91957750db63eb8c977f" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750699 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="92f449c478541d4fc040be370d8fbfc4918ffb370dcd3a1fef9f2aeed113524a" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750706 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="600192eff2ee82990dc4e3bfc25d3f3cdde9d6e9447251423860a8d59ffdffb8" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750713 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="e2ef352d28dc64cdc17f622c2a5ce4825d891748d60228a87b0234886b2fbd46" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750721 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="de82cace4517a41df59b73d9a505b6d9bc00d26b88a1425eae7b033bb480c311" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750727 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="de0a318ea00fadfa414f9b9cb4effedf4f8cbc78ffcf755bb0af2d1ee29bee87" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750733 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="419c0ad19d57e0d824fc3545acc3c8fa4dc2b5da26669ee7c69100b12f6cf8ad" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750739 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="02d59e38675cdde34dc486df647b1eff69092a995a977d61d28b1c5da1ad6b1e" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750746 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="b99dfd374a5ce1c13a934b2f873d212ff72dfe71b28341e2c047ff8d80e0c669" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750752 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="0854a4d87b476088ca363dd19e5424c00e3887a3cd94f9bb2324db18aa9630d0" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750759 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="be4e0386cd97f8aa803f6981bdc1c529197af63b34156666d8f8d7263d0d4f3d" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750400 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"695ff04f9aad2b63580222fb53407fe88ab246ea589b941267e537a78643ddbd"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750828 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"13bb352c512127bb19cb77c2407f696591efaeb94de88485c71ef25ef7efd32b"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750844 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"2078680d2bb75db1546bb97e0d090d85113a906cd7a2b21900140fb580300a81"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750853 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"b461410d2b38540e3971bd0c38fb4b2285b2b4bea2ad91957750db63eb8c977f"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750862 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"92f449c478541d4fc040be370d8fbfc4918ffb370dcd3a1fef9f2aeed113524a"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750873 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"600192eff2ee82990dc4e3bfc25d3f3cdde9d6e9447251423860a8d59ffdffb8"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750882 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"e2ef352d28dc64cdc17f622c2a5ce4825d891748d60228a87b0234886b2fbd46"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750891 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"de82cace4517a41df59b73d9a505b6d9bc00d26b88a1425eae7b033bb480c311"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750900 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"de0a318ea00fadfa414f9b9cb4effedf4f8cbc78ffcf755bb0af2d1ee29bee87"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750909 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"419c0ad19d57e0d824fc3545acc3c8fa4dc2b5da26669ee7c69100b12f6cf8ad"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750918 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"02d59e38675cdde34dc486df647b1eff69092a995a977d61d28b1c5da1ad6b1e"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750926 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"b99dfd374a5ce1c13a934b2f873d212ff72dfe71b28341e2c047ff8d80e0c669"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750935 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"0854a4d87b476088ca363dd19e5424c00e3887a3cd94f9bb2324db18aa9630d0"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.750943 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"be4e0386cd97f8aa803f6981bdc1c529197af63b34156666d8f8d7263d0d4f3d"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.752439 4706 generic.go:334] "Generic (PLEG): container finished" podID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerID="c7619136346823904b27331aeb39977b3b1020d65267db3982003f2469a6f40c" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.752482 4706 generic.go:334] "Generic (PLEG): container finished" podID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerID="9f23d04861cc7026fe4bf45c52c3a91306ee4fea16da518473dfc8059c16ef9c" exitCode=0 Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.752503 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" event={"ID":"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76","Type":"ContainerDied","Data":"c7619136346823904b27331aeb39977b3b1020d65267db3982003f2469a6f40c"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.752539 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" event={"ID":"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76","Type":"ContainerDied","Data":"9f23d04861cc7026fe4bf45c52c3a91306ee4fea16da518473dfc8059c16ef9c"} Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.784587 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="525e7753-926d-429b-bbe0-f45e95cf80c3" path="/var/lib/kubelet/pods/525e7753-926d-429b-bbe0-f45e95cf80c3/volumes" Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.800750 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.962676 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-config-data\") pod \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.962769 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trfrj\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-kube-api-access-trfrj\") pod \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.962824 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-run-httpd\") pod \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.962874 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") pod \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.962911 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-log-httpd\") pod \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\" (UID: \"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76\") " Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.963487 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.963641 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.968749 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.969348 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-kube-api-access-trfrj" (OuterVolumeSpecName: "kube-api-access-trfrj") pod "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76"). InnerVolumeSpecName "kube-api-access-trfrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:02 crc kubenswrapper[4706]: I1127 07:32:02.994399 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-config-data" (OuterVolumeSpecName: "config-data") pod "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" (UID: "64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.064403 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trfrj\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-kube-api-access-trfrj\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.064438 4706 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.064448 4706 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.064457 4706 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.064466 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.327100 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-fqrtx"] Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.338610 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-fqrtx"] Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.345234 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-m4zj7"] Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.358000 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-m4zj7"] Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.372557 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-59949565db-f2499"] Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.372765 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/keystone-59949565db-f2499" podUID="b67a1baa-0133-4b6d-b490-20ff47b1f90f" containerName="keystone-api" containerID="cri-o://5c0d5aeb797927517a277670205f5bba69b69629a355e0c670e0024c294bc4bd" gracePeriod=30 Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410120 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone01e7-account-delete-9hdrm"] Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410422 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410437 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410449 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410455 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410462 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410469 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410490 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2b34ae-f2f3-4571-bc32-3f002a955862" containerName="mariadb-account-delete" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410497 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2b34ae-f2f3-4571-bc32-3f002a955862" containerName="mariadb-account-delete" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410507 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-server" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410512 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-server" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410519 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410525 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410535 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410541 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410553 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410558 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410568 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410575 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410585 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410592 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: E1127 07:32:03.410601 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410608 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410728 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410743 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410752 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410762 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410769 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e779bc-2eae-44aa-890f-bb9c36aac432" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410776 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" containerName="proxy-server" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410785 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2b34ae-f2f3-4571-bc32-3f002a955862" containerName="mariadb-account-delete" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410795 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b99e235-cb57-43e7-a63e-2f177dc7549d" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410805 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-log" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410813 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4b7d2b-970b-410d-b11c-e11ac0b8a4bb" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.410822 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dfa1095-7cae-4bdd-9615-d9a447134b6f" containerName="glance-httpd" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.411266 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.425788 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone01e7-account-delete-9hdrm"] Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.569849 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h97fw\" (UniqueName: \"kubernetes.io/projected/b78277f0-fc55-4f1f-bf70-6522fa206106-kube-api-access-h97fw\") pod \"keystone01e7-account-delete-9hdrm\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.569955 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts\") pod \"keystone01e7-account-delete-9hdrm\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.671147 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts\") pod \"keystone01e7-account-delete-9hdrm\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.671312 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h97fw\" (UniqueName: \"kubernetes.io/projected/b78277f0-fc55-4f1f-bf70-6522fa206106-kube-api-access-h97fw\") pod \"keystone01e7-account-delete-9hdrm\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.672034 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts\") pod \"keystone01e7-account-delete-9hdrm\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.688491 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h97fw\" (UniqueName: \"kubernetes.io/projected/b78277f0-fc55-4f1f-bf70-6522fa206106-kube-api-access-h97fw\") pod \"keystone01e7-account-delete-9hdrm\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.733215 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.793460 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" event={"ID":"64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76","Type":"ContainerDied","Data":"fca707dfa183a719c29bd821db4190413d057073aa4d1266726fe8a84ee0fa4a"} Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.793575 4706 scope.go:117] "RemoveContainer" containerID="c7619136346823904b27331aeb39977b3b1020d65267db3982003f2469a6f40c" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.793582 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.835601 4706 scope.go:117] "RemoveContainer" containerID="9f23d04861cc7026fe4bf45c52c3a91306ee4fea16da518473dfc8059c16ef9c" Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.855093 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs"] Nov 27 07:32:03 crc kubenswrapper[4706]: I1127 07:32:03.860023 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jd2hs"] Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.016737 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.023773 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.029923 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.152628 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone01e7-account-delete-9hdrm"] Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.192689 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-2" podUID="170c6f78-b315-483f-a997-bfed203bacb7" containerName="galera" containerID="cri-o://9a73ed5960d9161255e97432f54ed2f8b667401423d038d0f7ed0aca9c65379d" gracePeriod=30 Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.643424 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.644016 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/memcached-0" podUID="8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" containerName="memcached" containerID="cri-o://7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6" gracePeriod=30 Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.787584 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0685d9b5-6dde-48e8-ac3d-4d247c5ecf06" path="/var/lib/kubelet/pods/0685d9b5-6dde-48e8-ac3d-4d247c5ecf06/volumes" Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.788333 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76" path="/var/lib/kubelet/pods/64c9fa38-8c1d-4a4c-8e7c-8656c53b8a76/volumes" Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.788891 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c827e889-9073-49b1-a264-bfa826083f54" path="/var/lib/kubelet/pods/c827e889-9073-49b1-a264-bfa826083f54/volumes" Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.812190 4706 generic.go:334] "Generic (PLEG): container finished" podID="170c6f78-b315-483f-a997-bfed203bacb7" containerID="9a73ed5960d9161255e97432f54ed2f8b667401423d038d0f7ed0aca9c65379d" exitCode=0 Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.812263 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"170c6f78-b315-483f-a997-bfed203bacb7","Type":"ContainerDied","Data":"9a73ed5960d9161255e97432f54ed2f8b667401423d038d0f7ed0aca9c65379d"} Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.813326 4706 generic.go:334] "Generic (PLEG): container finished" podID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerID="92e722e9fa65d0e5b136e205cc44a77914327b36a18a8ecfad9a812d0b7ab28c" exitCode=1 Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.813378 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" event={"ID":"b78277f0-fc55-4f1f-bf70-6522fa206106","Type":"ContainerDied","Data":"92e722e9fa65d0e5b136e205cc44a77914327b36a18a8ecfad9a812d0b7ab28c"} Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.813399 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" event={"ID":"b78277f0-fc55-4f1f-bf70-6522fa206106","Type":"ContainerStarted","Data":"d3941f9970e4855dfea8aa59bb2b8101549ebfd468351e95b2d032c177aaff70"} Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.813858 4706 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" secret="" err="secret \"galera-openstack-dockercfg-8kwnw\" not found" Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.813902 4706 scope.go:117] "RemoveContainer" containerID="92e722e9fa65d0e5b136e205cc44a77914327b36a18a8ecfad9a812d0b7ab28c" Nov 27 07:32:04 crc kubenswrapper[4706]: E1127 07:32:04.901711 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 27 07:32:04 crc kubenswrapper[4706]: E1127 07:32:04.901769 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts podName:b78277f0-fc55-4f1f-bf70-6522fa206106 nodeName:}" failed. No retries permitted until 2025-11-27 07:32:05.401750483 +0000 UTC m=+1409.291341303 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts") pod "keystone01e7-account-delete-9hdrm" (UID: "b78277f0-fc55-4f1f-bf70-6522fa206106") : configmap "openstack-scripts" not found Nov 27 07:32:04 crc kubenswrapper[4706]: I1127 07:32:04.973748 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.020444 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.103893 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-operator-scripts\") pod \"170c6f78-b315-483f-a997-bfed203bacb7\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.103966 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-kolla-config\") pod \"170c6f78-b315-483f-a997-bfed203bacb7\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104062 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"170c6f78-b315-483f-a997-bfed203bacb7\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104092 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/170c6f78-b315-483f-a997-bfed203bacb7-config-data-generated\") pod \"170c6f78-b315-483f-a997-bfed203bacb7\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104108 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-config-data-default\") pod \"170c6f78-b315-483f-a997-bfed203bacb7\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104160 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzj8b\" (UniqueName: \"kubernetes.io/projected/170c6f78-b315-483f-a997-bfed203bacb7-kube-api-access-pzj8b\") pod \"170c6f78-b315-483f-a997-bfed203bacb7\" (UID: \"170c6f78-b315-483f-a997-bfed203bacb7\") " Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104535 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/170c6f78-b315-483f-a997-bfed203bacb7-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "170c6f78-b315-483f-a997-bfed203bacb7" (UID: "170c6f78-b315-483f-a997-bfed203bacb7"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104731 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "170c6f78-b315-483f-a997-bfed203bacb7" (UID: "170c6f78-b315-483f-a997-bfed203bacb7"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104803 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "170c6f78-b315-483f-a997-bfed203bacb7" (UID: "170c6f78-b315-483f-a997-bfed203bacb7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.104826 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "170c6f78-b315-483f-a997-bfed203bacb7" (UID: "170c6f78-b315-483f-a997-bfed203bacb7"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.109583 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/170c6f78-b315-483f-a997-bfed203bacb7-kube-api-access-pzj8b" (OuterVolumeSpecName: "kube-api-access-pzj8b") pod "170c6f78-b315-483f-a997-bfed203bacb7" (UID: "170c6f78-b315-483f-a997-bfed203bacb7"). InnerVolumeSpecName "kube-api-access-pzj8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.113326 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "mysql-db") pod "170c6f78-b315-483f-a997-bfed203bacb7" (UID: "170c6f78-b315-483f-a997-bfed203bacb7"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.205941 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.206008 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.206053 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.206073 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/170c6f78-b315-483f-a997-bfed203bacb7-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.206093 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/170c6f78-b315-483f-a997-bfed203bacb7-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.206113 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzj8b\" (UniqueName: \"kubernetes.io/projected/170c6f78-b315-483f-a997-bfed203bacb7-kube-api-access-pzj8b\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.231488 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.307625 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:05 crc kubenswrapper[4706]: E1127 07:32:05.408906 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 27 07:32:05 crc kubenswrapper[4706]: E1127 07:32:05.409018 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts podName:b78277f0-fc55-4f1f-bf70-6522fa206106 nodeName:}" failed. No retries permitted until 2025-11-27 07:32:06.408994509 +0000 UTC m=+1410.298585319 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts") pod "keystone01e7-account-delete-9hdrm" (UID: "b78277f0-fc55-4f1f-bf70-6522fa206106") : configmap "openstack-scripts" not found Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.463583 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.844578 4706 generic.go:334] "Generic (PLEG): container finished" podID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerID="f9a4347e44959d07d2bd4203e0cb22e04e94a7777118038228631721940ca775" exitCode=1 Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.844644 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" event={"ID":"b78277f0-fc55-4f1f-bf70-6522fa206106","Type":"ContainerDied","Data":"f9a4347e44959d07d2bd4203e0cb22e04e94a7777118038228631721940ca775"} Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.845206 4706 scope.go:117] "RemoveContainer" containerID="92e722e9fa65d0e5b136e205cc44a77914327b36a18a8ecfad9a812d0b7ab28c" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.845382 4706 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" secret="" err="secret \"galera-openstack-dockercfg-8kwnw\" not found" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.845444 4706 scope.go:117] "RemoveContainer" containerID="f9a4347e44959d07d2bd4203e0cb22e04e94a7777118038228631721940ca775" Nov 27 07:32:05 crc kubenswrapper[4706]: E1127 07:32:05.845799 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystone01e7-account-delete-9hdrm_glance-kuttl-tests(b78277f0-fc55-4f1f-bf70-6522fa206106)\"" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.854418 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.857016 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"170c6f78-b315-483f-a997-bfed203bacb7","Type":"ContainerDied","Data":"f6efe971654605f901f8eccb5ba3479ed48fdbb7294614313744438cdef4c6a1"} Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.921186 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.921985 4706 scope.go:117] "RemoveContainer" containerID="9a73ed5960d9161255e97432f54ed2f8b667401423d038d0f7ed0aca9c65379d" Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.922897 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/rabbitmq-server-0" podUID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerName="rabbitmq" containerID="cri-o://cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2" gracePeriod=604800 Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.927596 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 27 07:32:05 crc kubenswrapper[4706]: I1127 07:32:05.958145 4706 scope.go:117] "RemoveContainer" containerID="83d37dbcc8861c7c72d29999f0b93082cab4a7ed666c6efadb361064699d6e49" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.171929 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-1" podUID="3047c710-a2e6-455a-951a-88971379f86f" containerName="galera" containerID="cri-o://db924923123328745d70c4e3e07509a6c0a28b588ec32e6d8cd509d486ad926b" gracePeriod=28 Nov 27 07:32:06 crc kubenswrapper[4706]: E1127 07:32:06.424983 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 27 07:32:06 crc kubenswrapper[4706]: E1127 07:32:06.425068 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts podName:b78277f0-fc55-4f1f-bf70-6522fa206106 nodeName:}" failed. No retries permitted until 2025-11-27 07:32:08.425050379 +0000 UTC m=+1412.314641189 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts") pod "keystone01e7-account-delete-9hdrm" (UID: "b78277f0-fc55-4f1f-bf70-6522fa206106") : configmap "openstack-scripts" not found Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.477867 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq"] Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.478102 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" podUID="3b7ef303-eaeb-4081-a2b9-44fbbc439974" containerName="manager" containerID="cri-o://ed594cd76f55eb035d4feb6cb65c42fca83b863966f11dd5bcf5295a464b64bf" gracePeriod=10 Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.774591 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.825277 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="170c6f78-b315-483f-a997-bfed203bacb7" path="/var/lib/kubelet/pods/170c6f78-b315-483f-a997-bfed203bacb7/volumes" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.832799 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-q6g28"] Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.832976 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-q6g28" podUID="b1511901-0bd6-44e5-a199-c99065ee8138" containerName="registry-server" containerID="cri-o://23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320" gracePeriod=30 Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.849685 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd"] Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.862517 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/ec4f141c67eeb9c1393444d4ccbcc6636e144397d5ef774a2057a621649rdqd"] Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.873985 4706 generic.go:334] "Generic (PLEG): container finished" podID="8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" containerID="7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6" exitCode=0 Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.874047 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51","Type":"ContainerDied","Data":"7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6"} Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.874079 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51","Type":"ContainerDied","Data":"9acfd22bd8ce2b32c15aaa5c7bfca9811fb708e043b816efdc02295ab7b683f0"} Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.874096 4706 scope.go:117] "RemoveContainer" containerID="7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.874203 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.878849 4706 generic.go:334] "Generic (PLEG): container finished" podID="b67a1baa-0133-4b6d-b490-20ff47b1f90f" containerID="5c0d5aeb797927517a277670205f5bba69b69629a355e0c670e0024c294bc4bd" exitCode=0 Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.878903 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-59949565db-f2499" event={"ID":"b67a1baa-0133-4b6d-b490-20ff47b1f90f","Type":"ContainerDied","Data":"5c0d5aeb797927517a277670205f5bba69b69629a355e0c670e0024c294bc4bd"} Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.881573 4706 generic.go:334] "Generic (PLEG): container finished" podID="3b7ef303-eaeb-4081-a2b9-44fbbc439974" containerID="ed594cd76f55eb035d4feb6cb65c42fca83b863966f11dd5bcf5295a464b64bf" exitCode=0 Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.881638 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" event={"ID":"3b7ef303-eaeb-4081-a2b9-44fbbc439974","Type":"ContainerDied","Data":"ed594cd76f55eb035d4feb6cb65c42fca83b863966f11dd5bcf5295a464b64bf"} Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.884121 4706 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" secret="" err="secret \"galera-openstack-dockercfg-8kwnw\" not found" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.884797 4706 scope.go:117] "RemoveContainer" containerID="f9a4347e44959d07d2bd4203e0cb22e04e94a7777118038228631721940ca775" Nov 27 07:32:06 crc kubenswrapper[4706]: E1127 07:32:06.884975 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystone01e7-account-delete-9hdrm_glance-kuttl-tests(b78277f0-fc55-4f1f-bf70-6522fa206106)\"" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.931082 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7kkb\" (UniqueName: \"kubernetes.io/projected/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kube-api-access-l7kkb\") pod \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.931132 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-config-data\") pod \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.931206 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kolla-config\") pod \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\" (UID: \"8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51\") " Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.932542 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" (UID: "8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.933012 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-config-data" (OuterVolumeSpecName: "config-data") pod "8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" (UID: "8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.936364 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kube-api-access-l7kkb" (OuterVolumeSpecName: "kube-api-access-l7kkb") pod "8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" (UID: "8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51"). InnerVolumeSpecName "kube-api-access-l7kkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.942340 4706 scope.go:117] "RemoveContainer" containerID="7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6" Nov 27 07:32:06 crc kubenswrapper[4706]: E1127 07:32:06.945587 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6\": container with ID starting with 7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6 not found: ID does not exist" containerID="7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.945633 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6"} err="failed to get container status \"7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6\": rpc error: code = NotFound desc = could not find container \"7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6\": container with ID starting with 7ae720d66a4244c1813cae34ff41fd65da7116c3dae9f65bc593d8c2c743c2b6 not found: ID does not exist" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.966325 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:32:06 crc kubenswrapper[4706]: I1127 07:32:06.999185 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.037366 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7kkb\" (UniqueName: \"kubernetes.io/projected/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kube-api-access-l7kkb\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.037419 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.037434 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141317 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-config-data\") pod \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141393 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcvjt\" (UniqueName: \"kubernetes.io/projected/3b7ef303-eaeb-4081-a2b9-44fbbc439974-kube-api-access-tcvjt\") pod \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141428 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-credential-keys\") pod \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141465 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-scripts\") pod \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141485 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp4dt\" (UniqueName: \"kubernetes.io/projected/b67a1baa-0133-4b6d-b490-20ff47b1f90f-kube-api-access-tp4dt\") pod \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141551 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-fernet-keys\") pod \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\" (UID: \"b67a1baa-0133-4b6d-b490-20ff47b1f90f\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141573 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-apiservice-cert\") pod \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.141609 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-webhook-cert\") pod \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\" (UID: \"3b7ef303-eaeb-4081-a2b9-44fbbc439974\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.146385 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "3b7ef303-eaeb-4081-a2b9-44fbbc439974" (UID: "3b7ef303-eaeb-4081-a2b9-44fbbc439974"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.168908 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "3b7ef303-eaeb-4081-a2b9-44fbbc439974" (UID: "3b7ef303-eaeb-4081-a2b9-44fbbc439974"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.168940 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-scripts" (OuterVolumeSpecName: "scripts") pod "b67a1baa-0133-4b6d-b490-20ff47b1f90f" (UID: "b67a1baa-0133-4b6d-b490-20ff47b1f90f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.169510 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b67a1baa-0133-4b6d-b490-20ff47b1f90f" (UID: "b67a1baa-0133-4b6d-b490-20ff47b1f90f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.169612 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b67a1baa-0133-4b6d-b490-20ff47b1f90f" (UID: "b67a1baa-0133-4b6d-b490-20ff47b1f90f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.170565 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67a1baa-0133-4b6d-b490-20ff47b1f90f-kube-api-access-tp4dt" (OuterVolumeSpecName: "kube-api-access-tp4dt") pod "b67a1baa-0133-4b6d-b490-20ff47b1f90f" (UID: "b67a1baa-0133-4b6d-b490-20ff47b1f90f"). InnerVolumeSpecName "kube-api-access-tp4dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.171312 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b7ef303-eaeb-4081-a2b9-44fbbc439974-kube-api-access-tcvjt" (OuterVolumeSpecName: "kube-api-access-tcvjt") pod "3b7ef303-eaeb-4081-a2b9-44fbbc439974" (UID: "3b7ef303-eaeb-4081-a2b9-44fbbc439974"). InnerVolumeSpecName "kube-api-access-tcvjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.203087 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-config-data" (OuterVolumeSpecName: "config-data") pod "b67a1baa-0133-4b6d-b490-20ff47b1f90f" (UID: "b67a1baa-0133-4b6d-b490-20ff47b1f90f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.224937 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.230003 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242845 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242874 4706 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242883 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcvjt\" (UniqueName: \"kubernetes.io/projected/3b7ef303-eaeb-4081-a2b9-44fbbc439974-kube-api-access-tcvjt\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242892 4706 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242902 4706 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242909 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp4dt\" (UniqueName: \"kubernetes.io/projected/b67a1baa-0133-4b6d-b490-20ff47b1f90f-kube-api-access-tp4dt\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242916 4706 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b67a1baa-0133-4b6d-b490-20ff47b1f90f-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.242924 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b7ef303-eaeb-4081-a2b9-44fbbc439974-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: E1127 07:32:07.326758 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cbc527a_c7f7_4ffe_b49a_c2262ef0ac51.slice/crio-9acfd22bd8ce2b32c15aaa5c7bfca9811fb708e043b816efdc02295ab7b683f0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cbc527a_c7f7_4ffe_b49a_c2262ef0ac51.slice\": RecentStats: unable to find data in memory cache]" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.337204 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.445873 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rfwv\" (UniqueName: \"kubernetes.io/projected/b1511901-0bd6-44e5-a199-c99065ee8138-kube-api-access-8rfwv\") pod \"b1511901-0bd6-44e5-a199-c99065ee8138\" (UID: \"b1511901-0bd6-44e5-a199-c99065ee8138\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.450118 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1511901-0bd6-44e5-a199-c99065ee8138-kube-api-access-8rfwv" (OuterVolumeSpecName: "kube-api-access-8rfwv") pod "b1511901-0bd6-44e5-a199-c99065ee8138" (UID: "b1511901-0bd6-44e5-a199-c99065ee8138"). InnerVolumeSpecName "kube-api-access-8rfwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.507390 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.548167 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rfwv\" (UniqueName: \"kubernetes.io/projected/b1511901-0bd6-44e5-a199-c99065ee8138-kube-api-access-8rfwv\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.649637 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53a204c4-9b3c-4d83-a87f-bf48abe46f77-erlang-cookie-secret\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.649684 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-confd\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.649854 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.649893 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28glv\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-kube-api-access-28glv\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.649925 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-erlang-cookie\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.649963 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53a204c4-9b3c-4d83-a87f-bf48abe46f77-plugins-conf\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.650000 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-plugins\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.650046 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53a204c4-9b3c-4d83-a87f-bf48abe46f77-pod-info\") pod \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\" (UID: \"53a204c4-9b3c-4d83-a87f-bf48abe46f77\") " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.650488 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.650786 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.650880 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53a204c4-9b3c-4d83-a87f-bf48abe46f77-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.653725 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/53a204c4-9b3c-4d83-a87f-bf48abe46f77-pod-info" (OuterVolumeSpecName: "pod-info") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.654082 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-kube-api-access-28glv" (OuterVolumeSpecName: "kube-api-access-28glv") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "kube-api-access-28glv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.654188 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53a204c4-9b3c-4d83-a87f-bf48abe46f77-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.663393 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2" (OuterVolumeSpecName: "persistence") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.710408 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "53a204c4-9b3c-4d83-a87f-bf48abe46f77" (UID: "53a204c4-9b3c-4d83-a87f-bf48abe46f77"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.751933 4706 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53a204c4-9b3c-4d83-a87f-bf48abe46f77-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.751962 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.751998 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\") on node \"crc\" " Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.752011 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28glv\" (UniqueName: \"kubernetes.io/projected/53a204c4-9b3c-4d83-a87f-bf48abe46f77-kube-api-access-28glv\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.752021 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.752029 4706 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53a204c4-9b3c-4d83-a87f-bf48abe46f77-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.752038 4706 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53a204c4-9b3c-4d83-a87f-bf48abe46f77-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.752046 4706 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53a204c4-9b3c-4d83-a87f-bf48abe46f77-pod-info\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.765421 4706 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.765575 4706 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2") on node "crc" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.853101 4706 reconciler_common.go:293] "Volume detached for volume \"pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9f5371a1-535b-4367-b789-5d6d1aea36d2\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.894960 4706 generic.go:334] "Generic (PLEG): container finished" podID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerID="cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2" exitCode=0 Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.895056 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.895135 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"53a204c4-9b3c-4d83-a87f-bf48abe46f77","Type":"ContainerDied","Data":"cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2"} Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.895198 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"53a204c4-9b3c-4d83-a87f-bf48abe46f77","Type":"ContainerDied","Data":"655d8fb3c393e0e5cdb0f05d5e3d64cf0a6b4496c84295ecd358435cd6ce6715"} Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.895251 4706 scope.go:117] "RemoveContainer" containerID="cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.899231 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-59949565db-f2499" event={"ID":"b67a1baa-0133-4b6d-b490-20ff47b1f90f","Type":"ContainerDied","Data":"c2347c5e3516f02865f03f40a4415dfe61cc7eb79e6ff9067ec74c932a31bb33"} Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.899296 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-59949565db-f2499" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.900924 4706 generic.go:334] "Generic (PLEG): container finished" podID="b1511901-0bd6-44e5-a199-c99065ee8138" containerID="23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320" exitCode=0 Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.900967 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-q6g28" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.900976 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-q6g28" event={"ID":"b1511901-0bd6-44e5-a199-c99065ee8138","Type":"ContainerDied","Data":"23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320"} Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.901051 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-q6g28" event={"ID":"b1511901-0bd6-44e5-a199-c99065ee8138","Type":"ContainerDied","Data":"cb8d0d0c67f1d975d677ac13c2788ffe661f3101019ee63df86c0dbe1bfce76a"} Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.904182 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" event={"ID":"3b7ef303-eaeb-4081-a2b9-44fbbc439974","Type":"ContainerDied","Data":"bd89b2d90d72f608cf31d55f3feb1cd09de14170c1db4940ee648d2efb9a345e"} Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.904189 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.908092 4706 generic.go:334] "Generic (PLEG): container finished" podID="3047c710-a2e6-455a-951a-88971379f86f" containerID="db924923123328745d70c4e3e07509a6c0a28b588ec32e6d8cd509d486ad926b" exitCode=0 Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.908120 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3047c710-a2e6-455a-951a-88971379f86f","Type":"ContainerDied","Data":"db924923123328745d70c4e3e07509a6c0a28b588ec32e6d8cd509d486ad926b"} Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.936023 4706 scope.go:117] "RemoveContainer" containerID="a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.952976 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-q6g28"] Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.960363 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-q6g28"] Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.961711 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.968742 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.974308 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.979778 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-59949565db-f2499"] Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.981934 4706 scope.go:117] "RemoveContainer" containerID="cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2" Nov 27 07:32:07 crc kubenswrapper[4706]: E1127 07:32:07.982356 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2\": container with ID starting with cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2 not found: ID does not exist" containerID="cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.982390 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2"} err="failed to get container status \"cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2\": rpc error: code = NotFound desc = could not find container \"cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2\": container with ID starting with cf612b6fdf79a2c507904a413bd2a50a81e5dffc85fcb1e0e88749cae467a2d2 not found: ID does not exist" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.982410 4706 scope.go:117] "RemoveContainer" containerID="a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f" Nov 27 07:32:07 crc kubenswrapper[4706]: E1127 07:32:07.982688 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f\": container with ID starting with a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f not found: ID does not exist" containerID="a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.982712 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f"} err="failed to get container status \"a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f\": rpc error: code = NotFound desc = could not find container \"a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f\": container with ID starting with a8a27a7babb06f412a8be0f9e0c07f7552db52663445163e216a4b14b9f2de4f not found: ID does not exist" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.982726 4706 scope.go:117] "RemoveContainer" containerID="5c0d5aeb797927517a277670205f5bba69b69629a355e0c670e0024c294bc4bd" Nov 27 07:32:07 crc kubenswrapper[4706]: I1127 07:32:07.987586 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-59949565db-f2499"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.014702 4706 scope.go:117] "RemoveContainer" containerID="23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.029002 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.032772 4706 scope.go:117] "RemoveContainer" containerID="23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.033147 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84798567c5-8dsrq"] Nov 27 07:32:08 crc kubenswrapper[4706]: E1127 07:32:08.033382 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320\": container with ID starting with 23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320 not found: ID does not exist" containerID="23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.033437 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320"} err="failed to get container status \"23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320\": rpc error: code = NotFound desc = could not find container \"23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320\": container with ID starting with 23de621b9ce2c4024e881ffbf91fa95fb83390aa0e9b8a4e20aebc2823ebe320 not found: ID does not exist" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.033480 4706 scope.go:117] "RemoveContainer" containerID="ed594cd76f55eb035d4feb6cb65c42fca83b863966f11dd5bcf5295a464b64bf" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.057590 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-kolla-config\") pod \"3047c710-a2e6-455a-951a-88971379f86f\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.057717 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-config-data-default\") pod \"3047c710-a2e6-455a-951a-88971379f86f\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.057959 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"3047c710-a2e6-455a-951a-88971379f86f\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058014 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "3047c710-a2e6-455a-951a-88971379f86f" (UID: "3047c710-a2e6-455a-951a-88971379f86f"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058101 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-operator-scripts\") pod \"3047c710-a2e6-455a-951a-88971379f86f\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058202 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "3047c710-a2e6-455a-951a-88971379f86f" (UID: "3047c710-a2e6-455a-951a-88971379f86f"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058313 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3047c710-a2e6-455a-951a-88971379f86f-config-data-generated\") pod \"3047c710-a2e6-455a-951a-88971379f86f\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058392 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72dzc\" (UniqueName: \"kubernetes.io/projected/3047c710-a2e6-455a-951a-88971379f86f-kube-api-access-72dzc\") pod \"3047c710-a2e6-455a-951a-88971379f86f\" (UID: \"3047c710-a2e6-455a-951a-88971379f86f\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058727 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058798 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058767 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3047c710-a2e6-455a-951a-88971379f86f" (UID: "3047c710-a2e6-455a-951a-88971379f86f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.058840 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3047c710-a2e6-455a-951a-88971379f86f-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "3047c710-a2e6-455a-951a-88971379f86f" (UID: "3047c710-a2e6-455a-951a-88971379f86f"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.062269 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3047c710-a2e6-455a-951a-88971379f86f-kube-api-access-72dzc" (OuterVolumeSpecName: "kube-api-access-72dzc") pod "3047c710-a2e6-455a-951a-88971379f86f" (UID: "3047c710-a2e6-455a-951a-88971379f86f"). InnerVolumeSpecName "kube-api-access-72dzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.070257 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "3047c710-a2e6-455a-951a-88971379f86f" (UID: "3047c710-a2e6-455a-951a-88971379f86f"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.160786 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.164341 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3047c710-a2e6-455a-951a-88971379f86f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.164399 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3047c710-a2e6-455a-951a-88971379f86f-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.164420 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72dzc\" (UniqueName: \"kubernetes.io/projected/3047c710-a2e6-455a-951a-88971379f86f-kube-api-access-72dzc\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.171864 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.190503 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-0" podUID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerName="galera" containerID="cri-o://d1104eea8593de00b4dfb366253def29206277bf5e1f33d94f3c0712811050fa" gracePeriod=26 Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.266428 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.434831 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-rwgn5"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.446319 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-rwgn5"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.462034 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.467272 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone01e7-account-delete-9hdrm"] Nov 27 07:32:08 crc kubenswrapper[4706]: E1127 07:32:08.469918 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 27 07:32:08 crc kubenswrapper[4706]: E1127 07:32:08.470047 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts podName:b78277f0-fc55-4f1f-bf70-6522fa206106 nodeName:}" failed. No retries permitted until 2025-11-27 07:32:12.470012113 +0000 UTC m=+1416.359602923 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts") pod "keystone01e7-account-delete-9hdrm" (UID: "b78277f0-fc55-4f1f-bf70-6522fa206106") : configmap "openstack-scripts" not found Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.471420 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-01e7-account-create-update-f6vfs"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.784517 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0917ef90-d580-44fd-b603-280fce1f2aec" path="/var/lib/kubelet/pods/0917ef90-d580-44fd-b603-280fce1f2aec/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.785391 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b7ef303-eaeb-4081-a2b9-44fbbc439974" path="/var/lib/kubelet/pods/3b7ef303-eaeb-4081-a2b9-44fbbc439974/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.786009 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" path="/var/lib/kubelet/pods/53a204c4-9b3c-4d83-a87f-bf48abe46f77/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.789525 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="680b9fbb-b235-4a8e-90d3-efbebbc3ba8f" path="/var/lib/kubelet/pods/680b9fbb-b235-4a8e-90d3-efbebbc3ba8f/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.789951 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" path="/var/lib/kubelet/pods/8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.790421 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1511901-0bd6-44e5-a199-c99065ee8138" path="/var/lib/kubelet/pods/b1511901-0bd6-44e5-a199-c99065ee8138/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.791615 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b613bb93-3d09-4cbb-9b7b-f0c9198f2412" path="/var/lib/kubelet/pods/b613bb93-3d09-4cbb-9b7b-f0c9198f2412/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.792141 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67a1baa-0133-4b6d-b490-20ff47b1f90f" path="/var/lib/kubelet/pods/b67a1baa-0133-4b6d-b490-20ff47b1f90f/volumes" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.806848 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.916166 4706 generic.go:334] "Generic (PLEG): container finished" podID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerID="d1104eea8593de00b4dfb366253def29206277bf5e1f33d94f3c0712811050fa" exitCode=0 Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.916255 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d","Type":"ContainerDied","Data":"d1104eea8593de00b4dfb366253def29206277bf5e1f33d94f3c0712811050fa"} Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.916295 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d","Type":"ContainerDied","Data":"7d8993f378a9ef0c8619418490c1b47b3d36cf1cab0b7ac7d1fd785962f65c80"} Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.916309 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d8993f378a9ef0c8619418490c1b47b3d36cf1cab0b7ac7d1fd785962f65c80" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.920149 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"3047c710-a2e6-455a-951a-88971379f86f","Type":"ContainerDied","Data":"1f839b1f75a6d582efeaa36446071fb0d0082f8612949d9c304e2d7de4939369"} Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.920177 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.920181 4706 scope.go:117] "RemoveContainer" containerID="db924923123328745d70c4e3e07509a6c0a28b588ec32e6d8cd509d486ad926b" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.921937 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" event={"ID":"b78277f0-fc55-4f1f-bf70-6522fa206106","Type":"ContainerDied","Data":"d3941f9970e4855dfea8aa59bb2b8101549ebfd468351e95b2d032c177aaff70"} Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.921975 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone01e7-account-delete-9hdrm" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.929924 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.939969 4706 scope.go:117] "RemoveContainer" containerID="65df9453f65098dc10938c2788c8a4b222dd248676351195b7bf545aec7cdf6e" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.940501 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.944918 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.962128 4706 scope.go:117] "RemoveContainer" containerID="f9a4347e44959d07d2bd4203e0cb22e04e94a7777118038228631721940ca775" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.977750 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h97fw\" (UniqueName: \"kubernetes.io/projected/b78277f0-fc55-4f1f-bf70-6522fa206106-kube-api-access-h97fw\") pod \"b78277f0-fc55-4f1f-bf70-6522fa206106\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.977897 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts\") pod \"b78277f0-fc55-4f1f-bf70-6522fa206106\" (UID: \"b78277f0-fc55-4f1f-bf70-6522fa206106\") " Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.979125 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b78277f0-fc55-4f1f-bf70-6522fa206106" (UID: "b78277f0-fc55-4f1f-bf70-6522fa206106"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:08 crc kubenswrapper[4706]: I1127 07:32:08.982912 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b78277f0-fc55-4f1f-bf70-6522fa206106-kube-api-access-h97fw" (OuterVolumeSpecName: "kube-api-access-h97fw") pod "b78277f0-fc55-4f1f-bf70-6522fa206106" (UID: "b78277f0-fc55-4f1f-bf70-6522fa206106"). InnerVolumeSpecName "kube-api-access-h97fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.079037 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-generated\") pod \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.079421 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.079442 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" (UID: "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.079502 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-default\") pod \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.079571 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-operator-scripts\") pod \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.079608 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kolla-config\") pod \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.079683 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzlr9\" (UniqueName: \"kubernetes.io/projected/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kube-api-access-jzlr9\") pod \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\" (UID: \"00a2cfb2-44cd-4fa1-a26a-f87e90b5165d\") " Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080089 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" (UID: "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080153 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" (UID: "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080623 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" (UID: "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080875 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080898 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b78277f0-fc55-4f1f-bf70-6522fa206106-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080928 4706 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080937 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h97fw\" (UniqueName: \"kubernetes.io/projected/b78277f0-fc55-4f1f-bf70-6522fa206106-kube-api-access-h97fw\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080945 4706 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.080953 4706 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.083109 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kube-api-access-jzlr9" (OuterVolumeSpecName: "kube-api-access-jzlr9") pod "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" (UID: "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d"). InnerVolumeSpecName "kube-api-access-jzlr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.088882 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" (UID: "00a2cfb2-44cd-4fa1-a26a-f87e90b5165d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.182340 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzlr9\" (UniqueName: \"kubernetes.io/projected/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d-kube-api-access-jzlr9\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.182468 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.210023 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.254258 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone01e7-account-delete-9hdrm"] Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.261906 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone01e7-account-delete-9hdrm"] Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.283844 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.936577 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.963170 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 27 07:32:09 crc kubenswrapper[4706]: I1127 07:32:09.968795 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.132031 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.132570 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" podUID="4e65f1dd-27d0-4445-b716-7e656f7c85c6" containerName="manager" containerID="cri-o://76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57" gracePeriod=10 Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.363755 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-wp5gg"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.364012 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-wp5gg" podUID="19bb9c9d-0e2c-443e-89c5-97987923c1f7" containerName="registry-server" containerID="cri-o://3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626" gracePeriod=30 Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.396195 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.402104 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bqg5k9"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.573471 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.709629 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-webhook-cert\") pod \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.709763 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-apiservice-cert\") pod \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.709805 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck4wh\" (UniqueName: \"kubernetes.io/projected/4e65f1dd-27d0-4445-b716-7e656f7c85c6-kube-api-access-ck4wh\") pod \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\" (UID: \"4e65f1dd-27d0-4445-b716-7e656f7c85c6\") " Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.714981 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e65f1dd-27d0-4445-b716-7e656f7c85c6-kube-api-access-ck4wh" (OuterVolumeSpecName: "kube-api-access-ck4wh") pod "4e65f1dd-27d0-4445-b716-7e656f7c85c6" (UID: "4e65f1dd-27d0-4445-b716-7e656f7c85c6"). InnerVolumeSpecName "kube-api-access-ck4wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.715039 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "4e65f1dd-27d0-4445-b716-7e656f7c85c6" (UID: "4e65f1dd-27d0-4445-b716-7e656f7c85c6"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.715116 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "4e65f1dd-27d0-4445-b716-7e656f7c85c6" (UID: "4e65f1dd-27d0-4445-b716-7e656f7c85c6"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.780559 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.785898 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" path="/var/lib/kubelet/pods/00a2cfb2-44cd-4fa1-a26a-f87e90b5165d/volumes" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.786983 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3047c710-a2e6-455a-951a-88971379f86f" path="/var/lib/kubelet/pods/3047c710-a2e6-455a-951a-88971379f86f/volumes" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.787644 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" path="/var/lib/kubelet/pods/b78277f0-fc55-4f1f-bf70-6522fa206106/volumes" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.788942 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf0b4f4f-059a-4c64-a75c-cffa8acd86ee" path="/var/lib/kubelet/pods/bf0b4f4f-059a-4c64-a75c-cffa8acd86ee/volumes" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.815253 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.815287 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck4wh\" (UniqueName: \"kubernetes.io/projected/4e65f1dd-27d0-4445-b716-7e656f7c85c6-kube-api-access-ck4wh\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.815299 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e65f1dd-27d0-4445-b716-7e656f7c85c6-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.916516 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r4md\" (UniqueName: \"kubernetes.io/projected/19bb9c9d-0e2c-443e-89c5-97987923c1f7-kube-api-access-5r4md\") pod \"19bb9c9d-0e2c-443e-89c5-97987923c1f7\" (UID: \"19bb9c9d-0e2c-443e-89c5-97987923c1f7\") " Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.919419 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19bb9c9d-0e2c-443e-89c5-97987923c1f7-kube-api-access-5r4md" (OuterVolumeSpecName: "kube-api-access-5r4md") pod "19bb9c9d-0e2c-443e-89c5-97987923c1f7" (UID: "19bb9c9d-0e2c-443e-89c5-97987923c1f7"). InnerVolumeSpecName "kube-api-access-5r4md". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.945006 4706 generic.go:334] "Generic (PLEG): container finished" podID="4e65f1dd-27d0-4445-b716-7e656f7c85c6" containerID="76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57" exitCode=0 Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.945072 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" event={"ID":"4e65f1dd-27d0-4445-b716-7e656f7c85c6","Type":"ContainerDied","Data":"76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57"} Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.945098 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" event={"ID":"4e65f1dd-27d0-4445-b716-7e656f7c85c6","Type":"ContainerDied","Data":"d54530a6e59ec41d30f64ff40e75fc5520b755e0635f6a5aee1de2a3c1fe6f1b"} Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.945116 4706 scope.go:117] "RemoveContainer" containerID="76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.945111 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.948013 4706 generic.go:334] "Generic (PLEG): container finished" podID="19bb9c9d-0e2c-443e-89c5-97987923c1f7" containerID="3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626" exitCode=0 Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.948041 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-wp5gg" event={"ID":"19bb9c9d-0e2c-443e-89c5-97987923c1f7","Type":"ContainerDied","Data":"3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626"} Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.948061 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-wp5gg" event={"ID":"19bb9c9d-0e2c-443e-89c5-97987923c1f7","Type":"ContainerDied","Data":"7d421fdcdf5bb347cc86539ed779ba32045f1e95d0b9548f56bf709a36f4476e"} Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.948079 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-wp5gg" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.963889 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.968597 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c4fb59c48-m6nmm"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.982988 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-wp5gg"] Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.984426 4706 scope.go:117] "RemoveContainer" containerID="76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57" Nov 27 07:32:10 crc kubenswrapper[4706]: E1127 07:32:10.985158 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57\": container with ID starting with 76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57 not found: ID does not exist" containerID="76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.985189 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57"} err="failed to get container status \"76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57\": rpc error: code = NotFound desc = could not find container \"76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57\": container with ID starting with 76244eb70720c1fdb4581c74309adcfe9b5c68aa011bc3f5c93cc56804d3fa57 not found: ID does not exist" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.985253 4706 scope.go:117] "RemoveContainer" containerID="3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626" Nov 27 07:32:10 crc kubenswrapper[4706]: I1127 07:32:10.988070 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-wp5gg"] Nov 27 07:32:11 crc kubenswrapper[4706]: I1127 07:32:11.012977 4706 scope.go:117] "RemoveContainer" containerID="3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626" Nov 27 07:32:11 crc kubenswrapper[4706]: E1127 07:32:11.013841 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626\": container with ID starting with 3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626 not found: ID does not exist" containerID="3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626" Nov 27 07:32:11 crc kubenswrapper[4706]: I1127 07:32:11.013871 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626"} err="failed to get container status \"3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626\": rpc error: code = NotFound desc = could not find container \"3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626\": container with ID starting with 3e145bc346695ec4f5b7427f6794514cae71af630bf16125e91514a4e1489626 not found: ID does not exist" Nov 27 07:32:11 crc kubenswrapper[4706]: I1127 07:32:11.020045 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r4md\" (UniqueName: \"kubernetes.io/projected/19bb9c9d-0e2c-443e-89c5-97987923c1f7-kube-api-access-5r4md\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:12 crc kubenswrapper[4706]: I1127 07:32:12.783333 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19bb9c9d-0e2c-443e-89c5-97987923c1f7" path="/var/lib/kubelet/pods/19bb9c9d-0e2c-443e-89c5-97987923c1f7/volumes" Nov 27 07:32:12 crc kubenswrapper[4706]: I1127 07:32:12.784486 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e65f1dd-27d0-4445-b716-7e656f7c85c6" path="/var/lib/kubelet/pods/4e65f1dd-27d0-4445-b716-7e656f7c85c6/volumes" Nov 27 07:32:13 crc kubenswrapper[4706]: I1127 07:32:13.742774 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf"] Nov 27 07:32:13 crc kubenswrapper[4706]: I1127 07:32:13.742983 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" podUID="dacd565c-fccf-46b1-a3c5-4f42366f29c6" containerName="manager" containerID="cri-o://f8ca8527391e1b47007e002d069ed9ed7cc7fa669ee15bb4bcbc67af39064d05" gracePeriod=10 Nov 27 07:32:13 crc kubenswrapper[4706]: I1127 07:32:13.978372 4706 generic.go:334] "Generic (PLEG): container finished" podID="dacd565c-fccf-46b1-a3c5-4f42366f29c6" containerID="f8ca8527391e1b47007e002d069ed9ed7cc7fa669ee15bb4bcbc67af39064d05" exitCode=0 Nov 27 07:32:13 crc kubenswrapper[4706]: I1127 07:32:13.978604 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" event={"ID":"dacd565c-fccf-46b1-a3c5-4f42366f29c6","Type":"ContainerDied","Data":"f8ca8527391e1b47007e002d069ed9ed7cc7fa669ee15bb4bcbc67af39064d05"} Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.001007 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-5slwj"] Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.001244 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-5slwj" podUID="82a85ca2-e74d-48ef-8fb4-19edc42cddac" containerName="registry-server" containerID="cri-o://57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573" gracePeriod=30 Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.025749 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx"] Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.040010 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3s56xx"] Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.181113 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.363838 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-apiservice-cert\") pod \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.363915 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-webhook-cert\") pod \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.363954 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv2lr\" (UniqueName: \"kubernetes.io/projected/dacd565c-fccf-46b1-a3c5-4f42366f29c6-kube-api-access-wv2lr\") pod \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\" (UID: \"dacd565c-fccf-46b1-a3c5-4f42366f29c6\") " Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.370413 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "dacd565c-fccf-46b1-a3c5-4f42366f29c6" (UID: "dacd565c-fccf-46b1-a3c5-4f42366f29c6"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.371382 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "dacd565c-fccf-46b1-a3c5-4f42366f29c6" (UID: "dacd565c-fccf-46b1-a3c5-4f42366f29c6"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.377465 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dacd565c-fccf-46b1-a3c5-4f42366f29c6-kube-api-access-wv2lr" (OuterVolumeSpecName: "kube-api-access-wv2lr") pod "dacd565c-fccf-46b1-a3c5-4f42366f29c6" (UID: "dacd565c-fccf-46b1-a3c5-4f42366f29c6"). InnerVolumeSpecName "kube-api-access-wv2lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.400186 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.465278 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv2lr\" (UniqueName: \"kubernetes.io/projected/dacd565c-fccf-46b1-a3c5-4f42366f29c6-kube-api-access-wv2lr\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.465362 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.465377 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dacd565c-fccf-46b1-a3c5-4f42366f29c6-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.566543 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx72s\" (UniqueName: \"kubernetes.io/projected/82a85ca2-e74d-48ef-8fb4-19edc42cddac-kube-api-access-tx72s\") pod \"82a85ca2-e74d-48ef-8fb4-19edc42cddac\" (UID: \"82a85ca2-e74d-48ef-8fb4-19edc42cddac\") " Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.569884 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82a85ca2-e74d-48ef-8fb4-19edc42cddac-kube-api-access-tx72s" (OuterVolumeSpecName: "kube-api-access-tx72s") pod "82a85ca2-e74d-48ef-8fb4-19edc42cddac" (UID: "82a85ca2-e74d-48ef-8fb4-19edc42cddac"). InnerVolumeSpecName "kube-api-access-tx72s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.668311 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx72s\" (UniqueName: \"kubernetes.io/projected/82a85ca2-e74d-48ef-8fb4-19edc42cddac-kube-api-access-tx72s\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.785398 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02868e73-1352-47bd-8fb4-8697bb9e2587" path="/var/lib/kubelet/pods/02868e73-1352-47bd-8fb4-8697bb9e2587/volumes" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.987617 4706 generic.go:334] "Generic (PLEG): container finished" podID="82a85ca2-e74d-48ef-8fb4-19edc42cddac" containerID="57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573" exitCode=0 Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.987718 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5slwj" event={"ID":"82a85ca2-e74d-48ef-8fb4-19edc42cddac","Type":"ContainerDied","Data":"57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573"} Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.987748 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5slwj" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.987784 4706 scope.go:117] "RemoveContainer" containerID="57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573" Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.987766 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5slwj" event={"ID":"82a85ca2-e74d-48ef-8fb4-19edc42cddac","Type":"ContainerDied","Data":"4074a2934ec3e43ae415d3f2165177e1e455a2e81cf71cb49c069d38dd866cae"} Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.990598 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" event={"ID":"dacd565c-fccf-46b1-a3c5-4f42366f29c6","Type":"ContainerDied","Data":"42e6b9a9c333256fca96fbb9a1e9a9e7d9209cd8b9797795d985512bef88073e"} Nov 27 07:32:14 crc kubenswrapper[4706]: I1127 07:32:14.990689 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf" Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.019413 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-5slwj"] Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.023554 4706 scope.go:117] "RemoveContainer" containerID="57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573" Nov 27 07:32:15 crc kubenswrapper[4706]: E1127 07:32:15.024537 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573\": container with ID starting with 57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573 not found: ID does not exist" containerID="57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573" Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.024576 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573"} err="failed to get container status \"57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573\": rpc error: code = NotFound desc = could not find container \"57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573\": container with ID starting with 57159ce98deb7522785f44a107972b403b1884cb883c8b605b92f519c0d2f573 not found: ID does not exist" Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.024604 4706 scope.go:117] "RemoveContainer" containerID="f8ca8527391e1b47007e002d069ed9ed7cc7fa669ee15bb4bcbc67af39064d05" Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.033176 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-5slwj"] Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.037444 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf"] Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.041825 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b75c8848-4vjhf"] Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.178304 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:32:15 crc kubenswrapper[4706]: I1127 07:32:15.178364 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.082417 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb"] Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.082616 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" podUID="636f33ce-c9d8-46ce-99e3-db7408c6e65b" containerName="operator" containerID="cri-o://b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844" gracePeriod=10 Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.417513 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-bh7ff"] Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.418017 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" podUID="c1a79656-0142-4194-bd13-90928f67a27e" containerName="registry-server" containerID="cri-o://f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b" gracePeriod=30 Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.438392 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx"] Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.443298 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lhcrx"] Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.561293 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.696726 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgljh\" (UniqueName: \"kubernetes.io/projected/636f33ce-c9d8-46ce-99e3-db7408c6e65b-kube-api-access-dgljh\") pod \"636f33ce-c9d8-46ce-99e3-db7408c6e65b\" (UID: \"636f33ce-c9d8-46ce-99e3-db7408c6e65b\") " Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.712478 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636f33ce-c9d8-46ce-99e3-db7408c6e65b-kube-api-access-dgljh" (OuterVolumeSpecName: "kube-api-access-dgljh") pod "636f33ce-c9d8-46ce-99e3-db7408c6e65b" (UID: "636f33ce-c9d8-46ce-99e3-db7408c6e65b"). InnerVolumeSpecName "kube-api-access-dgljh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.787007 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0371adf5-a190-4d98-90db-b83dcd65b5b9" path="/var/lib/kubelet/pods/0371adf5-a190-4d98-90db-b83dcd65b5b9/volumes" Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.787691 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82a85ca2-e74d-48ef-8fb4-19edc42cddac" path="/var/lib/kubelet/pods/82a85ca2-e74d-48ef-8fb4-19edc42cddac/volumes" Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.788164 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dacd565c-fccf-46b1-a3c5-4f42366f29c6" path="/var/lib/kubelet/pods/dacd565c-fccf-46b1-a3c5-4f42366f29c6/volumes" Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.797931 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgljh\" (UniqueName: \"kubernetes.io/projected/636f33ce-c9d8-46ce-99e3-db7408c6e65b-kube-api-access-dgljh\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:16 crc kubenswrapper[4706]: I1127 07:32:16.828769 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.000295 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngzqj\" (UniqueName: \"kubernetes.io/projected/c1a79656-0142-4194-bd13-90928f67a27e-kube-api-access-ngzqj\") pod \"c1a79656-0142-4194-bd13-90928f67a27e\" (UID: \"c1a79656-0142-4194-bd13-90928f67a27e\") " Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.004206 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a79656-0142-4194-bd13-90928f67a27e-kube-api-access-ngzqj" (OuterVolumeSpecName: "kube-api-access-ngzqj") pod "c1a79656-0142-4194-bd13-90928f67a27e" (UID: "c1a79656-0142-4194-bd13-90928f67a27e"). InnerVolumeSpecName "kube-api-access-ngzqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.007671 4706 generic.go:334] "Generic (PLEG): container finished" podID="c1a79656-0142-4194-bd13-90928f67a27e" containerID="f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b" exitCode=0 Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.007750 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" event={"ID":"c1a79656-0142-4194-bd13-90928f67a27e","Type":"ContainerDied","Data":"f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b"} Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.007780 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" event={"ID":"c1a79656-0142-4194-bd13-90928f67a27e","Type":"ContainerDied","Data":"c22969ec6cabf22da245f81038dacd1d4d694d3474fc331bdd48901e16a425f1"} Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.007802 4706 scope.go:117] "RemoveContainer" containerID="f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.008113 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-bh7ff" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.012052 4706 generic.go:334] "Generic (PLEG): container finished" podID="636f33ce-c9d8-46ce-99e3-db7408c6e65b" containerID="b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844" exitCode=0 Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.012084 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" event={"ID":"636f33ce-c9d8-46ce-99e3-db7408c6e65b","Type":"ContainerDied","Data":"b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844"} Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.012106 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" event={"ID":"636f33ce-c9d8-46ce-99e3-db7408c6e65b","Type":"ContainerDied","Data":"373e22adbc8ce2bfd7c5abaf2e21213401642df17d0bcbb2889b9b7e93589168"} Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.012157 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.025266 4706 scope.go:117] "RemoveContainer" containerID="f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b" Nov 27 07:32:17 crc kubenswrapper[4706]: E1127 07:32:17.025726 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b\": container with ID starting with f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b not found: ID does not exist" containerID="f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.025763 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b"} err="failed to get container status \"f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b\": rpc error: code = NotFound desc = could not find container \"f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b\": container with ID starting with f25716c09114e4dfa57fa30b8de2399950aecac4c4cd55304f1019be9f519c6b not found: ID does not exist" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.025788 4706 scope.go:117] "RemoveContainer" containerID="b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.033879 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb"] Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.038594 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-px4kb"] Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.050468 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-bh7ff"] Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.051317 4706 scope.go:117] "RemoveContainer" containerID="b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844" Nov 27 07:32:17 crc kubenswrapper[4706]: E1127 07:32:17.051920 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844\": container with ID starting with b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844 not found: ID does not exist" containerID="b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.051967 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844"} err="failed to get container status \"b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844\": rpc error: code = NotFound desc = could not find container \"b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844\": container with ID starting with b0826f80b11ba3c0ab598071ba305a5636887a787551108dff59dd5001ccb844 not found: ID does not exist" Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.057372 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-bh7ff"] Nov 27 07:32:17 crc kubenswrapper[4706]: I1127 07:32:17.102393 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngzqj\" (UniqueName: \"kubernetes.io/projected/c1a79656-0142-4194-bd13-90928f67a27e-kube-api-access-ngzqj\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:18 crc kubenswrapper[4706]: I1127 07:32:18.785972 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636f33ce-c9d8-46ce-99e3-db7408c6e65b" path="/var/lib/kubelet/pods/636f33ce-c9d8-46ce-99e3-db7408c6e65b/volumes" Nov 27 07:32:18 crc kubenswrapper[4706]: I1127 07:32:18.787444 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a79656-0142-4194-bd13-90928f67a27e" path="/var/lib/kubelet/pods/c1a79656-0142-4194-bd13-90928f67a27e/volumes" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.195658 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp"] Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.195926 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="manager" containerID="cri-o://86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c" gracePeriod=10 Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.196008 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="kube-rbac-proxy" containerID="cri-o://ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9" gracePeriod=10 Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.468937 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-ccm9g"] Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.469401 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-ccm9g" podUID="5b31be76-beca-407c-a1a9-db1e8e3f100e" containerName="registry-server" containerID="cri-o://e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c" gracePeriod=30 Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.500376 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr"] Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.512001 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d24xpr"] Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.667557 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.765797 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv894\" (UniqueName: \"kubernetes.io/projected/006027de-9ecc-4ff3-96a6-6918afb24789-kube-api-access-dv894\") pod \"006027de-9ecc-4ff3-96a6-6918afb24789\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.765918 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-webhook-cert\") pod \"006027de-9ecc-4ff3-96a6-6918afb24789\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.766635 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-apiservice-cert\") pod \"006027de-9ecc-4ff3-96a6-6918afb24789\" (UID: \"006027de-9ecc-4ff3-96a6-6918afb24789\") " Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.771024 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "006027de-9ecc-4ff3-96a6-6918afb24789" (UID: "006027de-9ecc-4ff3-96a6-6918afb24789"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.771681 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006027de-9ecc-4ff3-96a6-6918afb24789-kube-api-access-dv894" (OuterVolumeSpecName: "kube-api-access-dv894") pod "006027de-9ecc-4ff3-96a6-6918afb24789" (UID: "006027de-9ecc-4ff3-96a6-6918afb24789"). InnerVolumeSpecName "kube-api-access-dv894". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.771758 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "006027de-9ecc-4ff3-96a6-6918afb24789" (UID: "006027de-9ecc-4ff3-96a6-6918afb24789"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.826233 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.868835 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv894\" (UniqueName: \"kubernetes.io/projected/006027de-9ecc-4ff3-96a6-6918afb24789-kube-api-access-dv894\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.868877 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.868892 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/006027de-9ecc-4ff3-96a6-6918afb24789-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.969800 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2tbf\" (UniqueName: \"kubernetes.io/projected/5b31be76-beca-407c-a1a9-db1e8e3f100e-kube-api-access-s2tbf\") pod \"5b31be76-beca-407c-a1a9-db1e8e3f100e\" (UID: \"5b31be76-beca-407c-a1a9-db1e8e3f100e\") " Nov 27 07:32:21 crc kubenswrapper[4706]: I1127 07:32:21.977546 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b31be76-beca-407c-a1a9-db1e8e3f100e-kube-api-access-s2tbf" (OuterVolumeSpecName: "kube-api-access-s2tbf") pod "5b31be76-beca-407c-a1a9-db1e8e3f100e" (UID: "5b31be76-beca-407c-a1a9-db1e8e3f100e"). InnerVolumeSpecName "kube-api-access-s2tbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.063074 4706 generic.go:334] "Generic (PLEG): container finished" podID="5b31be76-beca-407c-a1a9-db1e8e3f100e" containerID="e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c" exitCode=0 Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.063151 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ccm9g" event={"ID":"5b31be76-beca-407c-a1a9-db1e8e3f100e","Type":"ContainerDied","Data":"e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c"} Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.063174 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ccm9g" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.063871 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ccm9g" event={"ID":"5b31be76-beca-407c-a1a9-db1e8e3f100e","Type":"ContainerDied","Data":"c400281098673ce784a72f82e0b72e27861f8fd325b8715ecaaf1e0435e9aa91"} Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.063878 4706 scope.go:117] "RemoveContainer" containerID="e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.069676 4706 generic.go:334] "Generic (PLEG): container finished" podID="006027de-9ecc-4ff3-96a6-6918afb24789" containerID="ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9" exitCode=0 Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.069807 4706 generic.go:334] "Generic (PLEG): container finished" podID="006027de-9ecc-4ff3-96a6-6918afb24789" containerID="86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c" exitCode=0 Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.069902 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" event={"ID":"006027de-9ecc-4ff3-96a6-6918afb24789","Type":"ContainerDied","Data":"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9"} Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.069999 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" event={"ID":"006027de-9ecc-4ff3-96a6-6918afb24789","Type":"ContainerDied","Data":"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c"} Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.070100 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" event={"ID":"006027de-9ecc-4ff3-96a6-6918afb24789","Type":"ContainerDied","Data":"d8d749dc5c2a268095ace272046c549d3b0231a1c95164d04b21b67440e88aae"} Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.070277 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.071443 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2tbf\" (UniqueName: \"kubernetes.io/projected/5b31be76-beca-407c-a1a9-db1e8e3f100e-kube-api-access-s2tbf\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.105023 4706 scope.go:117] "RemoveContainer" containerID="e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c" Nov 27 07:32:22 crc kubenswrapper[4706]: E1127 07:32:22.106355 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c\": container with ID starting with e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c not found: ID does not exist" containerID="e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.106413 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c"} err="failed to get container status \"e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c\": rpc error: code = NotFound desc = could not find container \"e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c\": container with ID starting with e42ed423c1a7085e6cd322c865a4c1179de6f04652598db1ec357647914abd0c not found: ID does not exist" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.106447 4706 scope.go:117] "RemoveContainer" containerID="ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.106820 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-ccm9g"] Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.112842 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-ccm9g"] Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.118407 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp"] Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.122049 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79755fc49-rb6vp"] Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.127188 4706 scope.go:117] "RemoveContainer" containerID="86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.143674 4706 scope.go:117] "RemoveContainer" containerID="ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9" Nov 27 07:32:22 crc kubenswrapper[4706]: E1127 07:32:22.144059 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9\": container with ID starting with ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9 not found: ID does not exist" containerID="ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.144093 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9"} err="failed to get container status \"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9\": rpc error: code = NotFound desc = could not find container \"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9\": container with ID starting with ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9 not found: ID does not exist" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.144116 4706 scope.go:117] "RemoveContainer" containerID="86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c" Nov 27 07:32:22 crc kubenswrapper[4706]: E1127 07:32:22.144472 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c\": container with ID starting with 86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c not found: ID does not exist" containerID="86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.144509 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c"} err="failed to get container status \"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c\": rpc error: code = NotFound desc = could not find container \"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c\": container with ID starting with 86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c not found: ID does not exist" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.144534 4706 scope.go:117] "RemoveContainer" containerID="ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.145038 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9"} err="failed to get container status \"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9\": rpc error: code = NotFound desc = could not find container \"ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9\": container with ID starting with ceb6798b8b17e2915724f2484a94101d8526896a32cdf9e72e1ad16397eca8d9 not found: ID does not exist" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.145064 4706 scope.go:117] "RemoveContainer" containerID="86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.145525 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c"} err="failed to get container status \"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c\": rpc error: code = NotFound desc = could not find container \"86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c\": container with ID starting with 86d552553f94e16fc1514bb191843741452c21536438be159d43bf13b71d4d1c not found: ID does not exist" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.784395 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" path="/var/lib/kubelet/pods/006027de-9ecc-4ff3-96a6-6918afb24789/volumes" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.785025 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b31be76-beca-407c-a1a9-db1e8e3f100e" path="/var/lib/kubelet/pods/5b31be76-beca-407c-a1a9-db1e8e3f100e/volumes" Nov 27 07:32:22 crc kubenswrapper[4706]: I1127 07:32:22.785643 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94169577-2199-446a-9d53-8473269fc045" path="/var/lib/kubelet/pods/94169577-2199-446a-9d53-8473269fc045/volumes" Nov 27 07:32:23 crc kubenswrapper[4706]: I1127 07:32:23.416305 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg"] Nov 27 07:32:23 crc kubenswrapper[4706]: I1127 07:32:23.416602 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" podUID="4b1548e8-8230-4cd3-9579-ecc704681612" containerName="manager" containerID="cri-o://2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5" gracePeriod=10 Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:23.765896 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-5s7bl"] Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:23.766409 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-5s7bl" podUID="a5eff20a-aef4-43d4-b5a6-779aa196bdb3" containerName="registry-server" containerID="cri-o://feaac72fbe9a0b58641ccab10d5f231dd867b0b07582a01647eb3e0e7a608832" gracePeriod=30 Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:23.814540 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg"] Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:23.816838 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvg7hg"] Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:23.932006 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.088167 4706 generic.go:334] "Generic (PLEG): container finished" podID="a5eff20a-aef4-43d4-b5a6-779aa196bdb3" containerID="feaac72fbe9a0b58641ccab10d5f231dd867b0b07582a01647eb3e0e7a608832" exitCode=0 Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.088250 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5s7bl" event={"ID":"a5eff20a-aef4-43d4-b5a6-779aa196bdb3","Type":"ContainerDied","Data":"feaac72fbe9a0b58641ccab10d5f231dd867b0b07582a01647eb3e0e7a608832"} Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.089423 4706 generic.go:334] "Generic (PLEG): container finished" podID="4b1548e8-8230-4cd3-9579-ecc704681612" containerID="2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5" exitCode=0 Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.089445 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" event={"ID":"4b1548e8-8230-4cd3-9579-ecc704681612","Type":"ContainerDied","Data":"2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5"} Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.089459 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" event={"ID":"4b1548e8-8230-4cd3-9579-ecc704681612","Type":"ContainerDied","Data":"82eb6db289851aac5237d2a75165d37947c3f697215405b469b14f9eba88e75d"} Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.089475 4706 scope.go:117] "RemoveContainer" containerID="2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.089551 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.098563 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7rjg\" (UniqueName: \"kubernetes.io/projected/4b1548e8-8230-4cd3-9579-ecc704681612-kube-api-access-c7rjg\") pod \"4b1548e8-8230-4cd3-9579-ecc704681612\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.098615 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-webhook-cert\") pod \"4b1548e8-8230-4cd3-9579-ecc704681612\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.098728 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-apiservice-cert\") pod \"4b1548e8-8230-4cd3-9579-ecc704681612\" (UID: \"4b1548e8-8230-4cd3-9579-ecc704681612\") " Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.103529 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b1548e8-8230-4cd3-9579-ecc704681612-kube-api-access-c7rjg" (OuterVolumeSpecName: "kube-api-access-c7rjg") pod "4b1548e8-8230-4cd3-9579-ecc704681612" (UID: "4b1548e8-8230-4cd3-9579-ecc704681612"). InnerVolumeSpecName "kube-api-access-c7rjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.103633 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "4b1548e8-8230-4cd3-9579-ecc704681612" (UID: "4b1548e8-8230-4cd3-9579-ecc704681612"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.107812 4706 scope.go:117] "RemoveContainer" containerID="2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5" Nov 27 07:32:24 crc kubenswrapper[4706]: E1127 07:32:24.108264 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5\": container with ID starting with 2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5 not found: ID does not exist" containerID="2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.108291 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5"} err="failed to get container status \"2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5\": rpc error: code = NotFound desc = could not find container \"2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5\": container with ID starting with 2e243e5cc839bbe2f0f5be80fd10518209177c3902d6b1534915c30ff3c38dc5 not found: ID does not exist" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.109341 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "4b1548e8-8230-4cd3-9579-ecc704681612" (UID: "4b1548e8-8230-4cd3-9579-ecc704681612"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.200150 4706 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.200188 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7rjg\" (UniqueName: \"kubernetes.io/projected/4b1548e8-8230-4cd3-9579-ecc704681612-kube-api-access-c7rjg\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.200204 4706 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b1548e8-8230-4cd3-9579-ecc704681612-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.419871 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg"] Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.432950 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-759795887-m7ctg"] Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.615359 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.706412 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kmng\" (UniqueName: \"kubernetes.io/projected/a5eff20a-aef4-43d4-b5a6-779aa196bdb3-kube-api-access-8kmng\") pod \"a5eff20a-aef4-43d4-b5a6-779aa196bdb3\" (UID: \"a5eff20a-aef4-43d4-b5a6-779aa196bdb3\") " Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.709240 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5eff20a-aef4-43d4-b5a6-779aa196bdb3-kube-api-access-8kmng" (OuterVolumeSpecName: "kube-api-access-8kmng") pod "a5eff20a-aef4-43d4-b5a6-779aa196bdb3" (UID: "a5eff20a-aef4-43d4-b5a6-779aa196bdb3"). InnerVolumeSpecName "kube-api-access-8kmng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.784961 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b1548e8-8230-4cd3-9579-ecc704681612" path="/var/lib/kubelet/pods/4b1548e8-8230-4cd3-9579-ecc704681612/volumes" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.786054 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b5ddba9-e78b-486e-9b93-a2ab98e0a09f" path="/var/lib/kubelet/pods/8b5ddba9-e78b-486e-9b93-a2ab98e0a09f/volumes" Nov 27 07:32:24 crc kubenswrapper[4706]: I1127 07:32:24.808689 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kmng\" (UniqueName: \"kubernetes.io/projected/a5eff20a-aef4-43d4-b5a6-779aa196bdb3-kube-api-access-8kmng\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:25 crc kubenswrapper[4706]: I1127 07:32:25.101338 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5s7bl" Nov 27 07:32:25 crc kubenswrapper[4706]: I1127 07:32:25.101347 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5s7bl" event={"ID":"a5eff20a-aef4-43d4-b5a6-779aa196bdb3","Type":"ContainerDied","Data":"f49f7303c56ca132c8f770e36263041388f5d231f489f2d7354601f4c064a76a"} Nov 27 07:32:25 crc kubenswrapper[4706]: I1127 07:32:25.101477 4706 scope.go:117] "RemoveContainer" containerID="feaac72fbe9a0b58641ccab10d5f231dd867b0b07582a01647eb3e0e7a608832" Nov 27 07:32:25 crc kubenswrapper[4706]: I1127 07:32:25.129446 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-5s7bl"] Nov 27 07:32:25 crc kubenswrapper[4706]: I1127 07:32:25.146591 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-5s7bl"] Nov 27 07:32:26 crc kubenswrapper[4706]: I1127 07:32:26.789764 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5eff20a-aef4-43d4-b5a6-779aa196bdb3" path="/var/lib/kubelet/pods/a5eff20a-aef4-43d4-b5a6-779aa196bdb3/volumes" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.183647 4706 generic.go:334] "Generic (PLEG): container finished" podID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerID="87334b52ad9d36330c1c6887ab6aefaab2f4a4d21ec015aff098d9b7df9b66e5" exitCode=137 Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.184293 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"87334b52ad9d36330c1c6887ab6aefaab2f4a4d21ec015aff098d9b7df9b66e5"} Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.472267 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.539691 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-cache\") pod \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.539838 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cszr\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-kube-api-access-5cszr\") pod \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.539858 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.539904 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-lock\") pod \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.539953 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") pod \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\" (UID: \"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd\") " Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.540499 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-lock" (OuterVolumeSpecName: "lock") pod "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.540830 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-cache" (OuterVolumeSpecName: "cache") pod "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.540988 4706 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-lock\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.541001 4706 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-cache\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.545907 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.546185 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-kube-api-access-5cszr" (OuterVolumeSpecName: "kube-api-access-5cszr") pod "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd"). InnerVolumeSpecName "kube-api-access-5cszr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.547454 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "swift") pod "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" (UID: "85c33b82-9ee0-4e79-82ee-f9de8a9bfefd"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.642687 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cszr\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-kube-api-access-5cszr\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.642746 4706 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.642757 4706 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.653953 4706 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 27 07:32:32 crc kubenswrapper[4706]: I1127 07:32:32.743990 4706 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.213098 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"85c33b82-9ee0-4e79-82ee-f9de8a9bfefd","Type":"ContainerDied","Data":"77b8887f2bd7f2407d49a3fd116c53505031d0d71c8ea57b6eddaeb41be8490f"} Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.214352 4706 scope.go:117] "RemoveContainer" containerID="87334b52ad9d36330c1c6887ab6aefaab2f4a4d21ec015aff098d9b7df9b66e5" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.213262 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.250077 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.256413 4706 scope.go:117] "RemoveContainer" containerID="695ff04f9aad2b63580222fb53407fe88ab246ea589b941267e537a78643ddbd" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.259590 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.287520 4706 scope.go:117] "RemoveContainer" containerID="13bb352c512127bb19cb77c2407f696591efaeb94de88485c71ef25ef7efd32b" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.309699 4706 scope.go:117] "RemoveContainer" containerID="2078680d2bb75db1546bb97e0d090d85113a906cd7a2b21900140fb580300a81" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.330045 4706 scope.go:117] "RemoveContainer" containerID="b461410d2b38540e3971bd0c38fb4b2285b2b4bea2ad91957750db63eb8c977f" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.370259 4706 scope.go:117] "RemoveContainer" containerID="92f449c478541d4fc040be370d8fbfc4918ffb370dcd3a1fef9f2aeed113524a" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.395592 4706 scope.go:117] "RemoveContainer" containerID="600192eff2ee82990dc4e3bfc25d3f3cdde9d6e9447251423860a8d59ffdffb8" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.426175 4706 scope.go:117] "RemoveContainer" containerID="e2ef352d28dc64cdc17f622c2a5ce4825d891748d60228a87b0234886b2fbd46" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.452167 4706 scope.go:117] "RemoveContainer" containerID="de82cace4517a41df59b73d9a505b6d9bc00d26b88a1425eae7b033bb480c311" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.471892 4706 scope.go:117] "RemoveContainer" containerID="de0a318ea00fadfa414f9b9cb4effedf4f8cbc78ffcf755bb0af2d1ee29bee87" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.491407 4706 scope.go:117] "RemoveContainer" containerID="419c0ad19d57e0d824fc3545acc3c8fa4dc2b5da26669ee7c69100b12f6cf8ad" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.518103 4706 scope.go:117] "RemoveContainer" containerID="02d59e38675cdde34dc486df647b1eff69092a995a977d61d28b1c5da1ad6b1e" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.539941 4706 scope.go:117] "RemoveContainer" containerID="b99dfd374a5ce1c13a934b2f873d212ff72dfe71b28341e2c047ff8d80e0c669" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.564300 4706 scope.go:117] "RemoveContainer" containerID="0854a4d87b476088ca363dd19e5424c00e3887a3cd94f9bb2324db18aa9630d0" Nov 27 07:32:33 crc kubenswrapper[4706]: I1127 07:32:33.583241 4706 scope.go:117] "RemoveContainer" containerID="be4e0386cd97f8aa803f6981bdc1c529197af63b34156666d8f8d7263d0d4f3d" Nov 27 07:32:34 crc kubenswrapper[4706]: I1127 07:32:34.793966 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" path="/var/lib/kubelet/pods/85c33b82-9ee0-4e79-82ee-f9de8a9bfefd/volumes" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285044 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jrcsw/must-gather-d65pf"] Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285599 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19bb9c9d-0e2c-443e-89c5-97987923c1f7" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285614 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="19bb9c9d-0e2c-443e-89c5-97987923c1f7" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285623 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285629 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285640 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerName="mariadb-account-delete" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285647 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerName="mariadb-account-delete" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285665 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" containerName="memcached" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285672 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" containerName="memcached" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285681 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285688 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285701 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5eff20a-aef4-43d4-b5a6-779aa196bdb3" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285709 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5eff20a-aef4-43d4-b5a6-779aa196bdb3" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285716 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerName="setup-container" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285731 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerName="setup-container" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285740 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-reaper" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285746 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-reaper" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285754 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285760 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285767 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1511901-0bd6-44e5-a199-c99065ee8138" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285774 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1511901-0bd6-44e5-a199-c99065ee8138" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285786 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285794 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285802 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1548e8-8230-4cd3-9579-ecc704681612" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285808 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1548e8-8230-4cd3-9579-ecc704681612" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285818 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636f33ce-c9d8-46ce-99e3-db7408c6e65b" containerName="operator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285825 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="636f33ce-c9d8-46ce-99e3-db7408c6e65b" containerName="operator" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285833 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-updater" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285839 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-updater" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285848 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="kube-rbac-proxy" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285854 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="kube-rbac-proxy" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285861 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285867 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285879 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285887 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285898 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-updater" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285904 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-updater" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285911 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3047c710-a2e6-455a-951a-88971379f86f" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285917 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3047c710-a2e6-455a-951a-88971379f86f" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285924 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="swift-recon-cron" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285930 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="swift-recon-cron" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285939 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e65f1dd-27d0-4445-b716-7e656f7c85c6" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285945 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e65f1dd-27d0-4445-b716-7e656f7c85c6" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285952 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285958 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285968 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a85ca2-e74d-48ef-8fb4-19edc42cddac" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285974 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a85ca2-e74d-48ef-8fb4-19edc42cddac" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285981 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a79656-0142-4194-bd13-90928f67a27e" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.285986 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a79656-0142-4194-bd13-90928f67a27e" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.285996 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="rsync" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286002 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="rsync" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286013 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b31be76-beca-407c-a1a9-db1e8e3f100e" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286019 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b31be76-beca-407c-a1a9-db1e8e3f100e" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286028 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286034 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286042 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="170c6f78-b315-483f-a997-bfed203bacb7" containerName="mysql-bootstrap" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286048 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="170c6f78-b315-483f-a997-bfed203bacb7" containerName="mysql-bootstrap" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286056 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7ef303-eaeb-4081-a2b9-44fbbc439974" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286062 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7ef303-eaeb-4081-a2b9-44fbbc439974" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286070 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-expirer" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286077 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-expirer" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286087 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3047c710-a2e6-455a-951a-88971379f86f" containerName="mysql-bootstrap" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286094 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="3047c710-a2e6-455a-951a-88971379f86f" containerName="mysql-bootstrap" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286102 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dacd565c-fccf-46b1-a3c5-4f42366f29c6" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286108 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="dacd565c-fccf-46b1-a3c5-4f42366f29c6" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286114 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286119 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286126 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67a1baa-0133-4b6d-b490-20ff47b1f90f" containerName="keystone-api" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286132 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67a1baa-0133-4b6d-b490-20ff47b1f90f" containerName="keystone-api" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286140 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="170c6f78-b315-483f-a997-bfed203bacb7" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286145 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="170c6f78-b315-483f-a997-bfed203bacb7" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286153 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286159 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286167 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerName="mysql-bootstrap" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286173 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerName="mysql-bootstrap" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286180 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerName="rabbitmq" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286185 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerName="rabbitmq" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286193 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286198 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286333 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-updater" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286344 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b1548e8-8230-4cd3-9579-ecc704681612" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286350 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="636f33ce-c9d8-46ce-99e3-db7408c6e65b" containerName="operator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286361 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286366 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="rsync" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286376 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5eff20a-aef4-43d4-b5a6-779aa196bdb3" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286384 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-expirer" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286391 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b7ef303-eaeb-4081-a2b9-44fbbc439974" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286397 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b31be76-beca-407c-a1a9-db1e8e3f100e" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286404 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a2cfb2-44cd-4fa1-a26a-f87e90b5165d" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286413 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286422 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286428 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286436 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="19bb9c9d-0e2c-443e-89c5-97987923c1f7" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286442 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286449 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286456 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1511901-0bd6-44e5-a199-c99065ee8138" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286462 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbc527a-c7f7-4ffe-b49a-c2262ef0ac51" containerName="memcached" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286468 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="82a85ca2-e74d-48ef-8fb4-19edc42cddac" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286473 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67a1baa-0133-4b6d-b490-20ff47b1f90f" containerName="keystone-api" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286480 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="container-updater" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286488 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-replicator" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286496 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="swift-recon-cron" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286504 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e65f1dd-27d0-4445-b716-7e656f7c85c6" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286512 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286518 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="object-auditor" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286526 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286534 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="006027de-9ecc-4ff3-96a6-6918afb24789" containerName="kube-rbac-proxy" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286541 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a204c4-9b3c-4d83-a87f-bf48abe46f77" containerName="rabbitmq" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286548 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a79656-0142-4194-bd13-90928f67a27e" containerName="registry-server" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286555 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c33b82-9ee0-4e79-82ee-f9de8a9bfefd" containerName="account-reaper" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286562 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerName="mariadb-account-delete" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286570 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="170c6f78-b315-483f-a997-bfed203bacb7" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286577 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="3047c710-a2e6-455a-951a-88971379f86f" containerName="galera" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286582 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="dacd565c-fccf-46b1-a3c5-4f42366f29c6" containerName="manager" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286589 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerName="mariadb-account-delete" Nov 27 07:32:37 crc kubenswrapper[4706]: E1127 07:32:37.286672 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerName="mariadb-account-delete" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.286679 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78277f0-fc55-4f1f-bf70-6522fa206106" containerName="mariadb-account-delete" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.287114 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.290304 4706 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jrcsw"/"default-dockercfg-5b4pz" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.290858 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jrcsw"/"openshift-service-ca.crt" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.291084 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jrcsw"/"kube-root-ca.crt" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.312989 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/abf4bcc1-8309-4b72-acd7-7b246a41eff6-must-gather-output\") pod \"must-gather-d65pf\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.315742 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxmdh\" (UniqueName: \"kubernetes.io/projected/abf4bcc1-8309-4b72-acd7-7b246a41eff6-kube-api-access-mxmdh\") pod \"must-gather-d65pf\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.315650 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jrcsw/must-gather-d65pf"] Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.417047 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/abf4bcc1-8309-4b72-acd7-7b246a41eff6-must-gather-output\") pod \"must-gather-d65pf\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.417087 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxmdh\" (UniqueName: \"kubernetes.io/projected/abf4bcc1-8309-4b72-acd7-7b246a41eff6-kube-api-access-mxmdh\") pod \"must-gather-d65pf\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.417576 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/abf4bcc1-8309-4b72-acd7-7b246a41eff6-must-gather-output\") pod \"must-gather-d65pf\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.438892 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxmdh\" (UniqueName: \"kubernetes.io/projected/abf4bcc1-8309-4b72-acd7-7b246a41eff6-kube-api-access-mxmdh\") pod \"must-gather-d65pf\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.605105 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.827499 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jrcsw/must-gather-d65pf"] Nov 27 07:32:37 crc kubenswrapper[4706]: I1127 07:32:37.875992 4706 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.242462 4706 scope.go:117] "RemoveContainer" containerID="df440ed18630a9f558e95670a1d131f78d1180569d871106239545490b467850" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.270739 4706 scope.go:117] "RemoveContainer" containerID="3c7cb477052ba49d8ac443222892feee698a2a40cb1a09f9c2f5731a541c1c07" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.292655 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jrcsw/must-gather-d65pf" event={"ID":"abf4bcc1-8309-4b72-acd7-7b246a41eff6","Type":"ContainerStarted","Data":"1172ce97c0adc493e11b8aa8ccef5a5c4a3355b9b526b9749507e41b8c20ab8f"} Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.310861 4706 scope.go:117] "RemoveContainer" containerID="2e9851ea626c3613fbaab116322ef135607b29f3b73e1c04c9b421e6edc66eab" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.339921 4706 scope.go:117] "RemoveContainer" containerID="3b4a2bd377e8eb5e0c81fe64c31c3c023dbb60de56a4ccc0dffa6544b8e63cb7" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.366014 4706 scope.go:117] "RemoveContainer" containerID="e2af845e375f59819a9d1bfe3d25bec314670153c8d2c8136ec610faec8e64ef" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.390538 4706 scope.go:117] "RemoveContainer" containerID="d1104eea8593de00b4dfb366253def29206277bf5e1f33d94f3c0712811050fa" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.440079 4706 scope.go:117] "RemoveContainer" containerID="7340b04b532e0329384704ffac9894d728fbe34a4cc96e8282e264d9a7a5e182" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.495069 4706 scope.go:117] "RemoveContainer" containerID="886b0fc2c88748cfd1ccdec5fc7b36fbdababc8eab452e815f5485e83223ee94" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.513423 4706 scope.go:117] "RemoveContainer" containerID="71c31a5f98996cc2836816887d95d7c8ebb6f2af0069aead745b5894d3b87270" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.554811 4706 scope.go:117] "RemoveContainer" containerID="9a1d43cb43cde6762ef8a3e95bdd5e4931e6273c991fc9bdd4df634ac90a3056" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.574923 4706 scope.go:117] "RemoveContainer" containerID="fba8bc694391c53a203d03f959f74bbd5575b1c342b6a2a1e47d4435cba08038" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.602022 4706 scope.go:117] "RemoveContainer" containerID="4a52d9d607e438f583ea3517ffab5bd0bc362f7711bb9bbb1fbb1f90aeebd843" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.644273 4706 scope.go:117] "RemoveContainer" containerID="090f20d74cd50b5f29e6c1f1d01ea1ccbc390f32108f3bce77c71c2285bcae32" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.663858 4706 scope.go:117] "RemoveContainer" containerID="2978ff29f5f5f9fc17133dae2cad42b5d07af45f577880c3a292eb34d02a54c3" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.682904 4706 scope.go:117] "RemoveContainer" containerID="a5ea7b183ead5b7612c935973b11381149d3cd0f87a35c1ea1ba8cec6d2c884d" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.704093 4706 scope.go:117] "RemoveContainer" containerID="dd4684d059a4f4303d5e038b480569409a65ff70f6401f9cafdad71df6843d5a" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.725648 4706 scope.go:117] "RemoveContainer" containerID="6fb01b105045e54c0460e23063053254e968a28290766f6c9c688dc0c5ec831b" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.763914 4706 scope.go:117] "RemoveContainer" containerID="fa2a6b2b83c7e6262e1e8885b82b53deb1a8f3687cc2e5205bd39b03c2dd89ce" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.786727 4706 scope.go:117] "RemoveContainer" containerID="3eb746183d3299d176023504f7c1a7ecfdfccb4456c29386fad43657c43cd0fa" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.804888 4706 scope.go:117] "RemoveContainer" containerID="a9b8ab89a86be81768bb0ca0e264236bebf22fb19be64499857af94a9d8d3032" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.822059 4706 scope.go:117] "RemoveContainer" containerID="92288fb3d75479196d597f61f708b5af55c44b0e13d806bd33700507b20cf72f" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.842338 4706 scope.go:117] "RemoveContainer" containerID="8e3901e1155389115253f4ec48b1d9d325f803566fd819e9b1ceb383df2d3d3b" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.885868 4706 scope.go:117] "RemoveContainer" containerID="f2cbe649faeed115dff4c5e03f81d485700d3dfd7addd47ff9fdb7a522a5b977" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.902860 4706 scope.go:117] "RemoveContainer" containerID="98c4fb18458a1412f153c1aafd7ddb8c730c571c8130eeec548811699554f7b4" Nov 27 07:32:38 crc kubenswrapper[4706]: I1127 07:32:38.918146 4706 scope.go:117] "RemoveContainer" containerID="873eb2e7db7b0b3a8f43e995cf3a2596e3f66e605538ddc07cd169bf4cf99bc3" Nov 27 07:32:42 crc kubenswrapper[4706]: I1127 07:32:42.321680 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jrcsw/must-gather-d65pf" event={"ID":"abf4bcc1-8309-4b72-acd7-7b246a41eff6","Type":"ContainerStarted","Data":"c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d"} Nov 27 07:32:42 crc kubenswrapper[4706]: I1127 07:32:42.322003 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jrcsw/must-gather-d65pf" event={"ID":"abf4bcc1-8309-4b72-acd7-7b246a41eff6","Type":"ContainerStarted","Data":"8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603"} Nov 27 07:32:42 crc kubenswrapper[4706]: I1127 07:32:42.340717 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jrcsw/must-gather-d65pf" podStartSLOduration=1.670059804 podStartE2EDuration="5.340698976s" podCreationTimestamp="2025-11-27 07:32:37 +0000 UTC" firstStartedPulling="2025-11-27 07:32:37.875756504 +0000 UTC m=+1441.765347314" lastFinishedPulling="2025-11-27 07:32:41.546395686 +0000 UTC m=+1445.435986486" observedRunningTime="2025-11-27 07:32:42.335101797 +0000 UTC m=+1446.224692617" watchObservedRunningTime="2025-11-27 07:32:42.340698976 +0000 UTC m=+1446.230289776" Nov 27 07:32:45 crc kubenswrapper[4706]: I1127 07:32:45.178148 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:32:45 crc kubenswrapper[4706]: I1127 07:32:45.178278 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:33:12 crc kubenswrapper[4706]: I1127 07:33:12.861911 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/util/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.054011 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/pull/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.057512 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/util/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.086588 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/pull/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.214721 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/util/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.219984 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/pull/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.268342 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/extract/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.371209 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54fdcb5fdb-dn97q_0961e9eb-dadf-4e6a-bbbb-34cd06d1b683/manager/0.log" Nov 27 07:33:13 crc kubenswrapper[4706]: I1127 07:33:13.413503 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-jnvft_7e5fd17c-d97a-4826-bd68-cd8480f1363d/registry-server/0.log" Nov 27 07:33:14 crc kubenswrapper[4706]: E1127 07:33:14.843995 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:33:14 crc kubenswrapper[4706]: E1127 07:33:14.844080 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:15.344056836 +0000 UTC m=+1479.233647646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:33:14 crc kubenswrapper[4706]: E1127 07:33:14.844014 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:33:14 crc kubenswrapper[4706]: E1127 07:33:14.844175 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:15.344156839 +0000 UTC m=+1479.233747649 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.178038 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.178129 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.178196 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.179047 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.179146 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" gracePeriod=600 Nov 27 07:33:15 crc kubenswrapper[4706]: E1127 07:33:15.300951 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:33:15 crc kubenswrapper[4706]: E1127 07:33:15.352531 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:33:15 crc kubenswrapper[4706]: E1127 07:33:15.352624 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:16.352604377 +0000 UTC m=+1480.242195187 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:33:15 crc kubenswrapper[4706]: E1127 07:33:15.352627 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:33:15 crc kubenswrapper[4706]: E1127 07:33:15.352720 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:16.352692779 +0000 UTC m=+1480.242283659 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.533922 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" exitCode=0 Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.533995 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900"} Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.534053 4706 scope.go:117] "RemoveContainer" containerID="6fa6ce74ad58cab1c7d76467a227f52ada0afc016b50f3d7c2b5893e9773b0ee" Nov 27 07:33:15 crc kubenswrapper[4706]: I1127 07:33:15.534624 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:33:15 crc kubenswrapper[4706]: E1127 07:33:15.534855 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:33:16 crc kubenswrapper[4706]: E1127 07:33:16.384215 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:33:16 crc kubenswrapper[4706]: E1127 07:33:16.384266 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:33:16 crc kubenswrapper[4706]: E1127 07:33:16.384303 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:18.384287784 +0000 UTC m=+1482.273878594 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:33:16 crc kubenswrapper[4706]: E1127 07:33:16.384386 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:18.384356906 +0000 UTC m=+1482.273947746 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:33:18 crc kubenswrapper[4706]: E1127 07:33:18.407977 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:33:18 crc kubenswrapper[4706]: E1127 07:33:18.408476 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:22.408450715 +0000 UTC m=+1486.298041555 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:33:18 crc kubenswrapper[4706]: E1127 07:33:18.408095 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:33:18 crc kubenswrapper[4706]: E1127 07:33:18.408539 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:22.408525648 +0000 UTC m=+1486.298116468 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:33:22 crc kubenswrapper[4706]: E1127 07:33:22.473729 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:33:22 crc kubenswrapper[4706]: E1127 07:33:22.474141 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:30.474117449 +0000 UTC m=+1494.363708279 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:33:22 crc kubenswrapper[4706]: E1127 07:33:22.473794 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:33:22 crc kubenswrapper[4706]: E1127 07:33:22.474369 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:30.474315275 +0000 UTC m=+1494.363906125 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:33:25 crc kubenswrapper[4706]: I1127 07:33:25.765824 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-c878d_dd986c7f-e3ae-4d24-8815-c38d546168e7/control-plane-machine-set-operator/0.log" Nov 27 07:33:25 crc kubenswrapper[4706]: I1127 07:33:25.895178 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r6h9j_0b4e82db-e890-4dc6-a6f0-441d78878459/kube-rbac-proxy/0.log" Nov 27 07:33:25 crc kubenswrapper[4706]: I1127 07:33:25.949827 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r6h9j_0b4e82db-e890-4dc6-a6f0-441d78878459/machine-api-operator/0.log" Nov 27 07:33:30 crc kubenswrapper[4706]: E1127 07:33:30.570774 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:33:30 crc kubenswrapper[4706]: E1127 07:33:30.570844 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:46.570832095 +0000 UTC m=+1510.460422905 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:33:30 crc kubenswrapper[4706]: E1127 07:33:30.570864 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:33:30 crc kubenswrapper[4706]: E1127 07:33:30.570976 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:33:46.570958269 +0000 UTC m=+1510.460549079 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:33:30 crc kubenswrapper[4706]: I1127 07:33:30.776414 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:33:30 crc kubenswrapper[4706]: E1127 07:33:30.776601 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.343710 4706 scope.go:117] "RemoveContainer" containerID="12fa7d40f5a433302eddef5bd9fb154643de88e96eb951204d94bd7482fac427" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.382276 4706 scope.go:117] "RemoveContainer" containerID="233302eba3934529a083f1c8281a7359f350ca8cfdb0961953fe75c70dbf51ae" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.435232 4706 scope.go:117] "RemoveContainer" containerID="d13c12fb2ce58f6e2af465403ff0a52ce1dd6f62137f267bbcec4156b3b602b3" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.458373 4706 scope.go:117] "RemoveContainer" containerID="3e278089af7f8ca36bd2f8e0b74d74720499b1eac07ce3f65f93eb82f4a59b4c" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.484868 4706 scope.go:117] "RemoveContainer" containerID="d3830433d5c8e272ec09cf93094e448bf53b16a4d79969c062f25f5940a4e0f6" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.509447 4706 scope.go:117] "RemoveContainer" containerID="b8bddeef39708a438e79d5792543906f80ab1d0ef1cf6d97d6def7030e1fddc5" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.534316 4706 scope.go:117] "RemoveContainer" containerID="ab5487f54e864099c15b01429a9a5faeffb0301a91c0bf59f0b9441b7143d40a" Nov 27 07:33:39 crc kubenswrapper[4706]: I1127 07:33:39.549558 4706 scope.go:117] "RemoveContainer" containerID="b13b57ab0e5f398cb514f76ffe3b76d52d898f85bb3464ae1c021ac4f13b27a2" Nov 27 07:33:40 crc kubenswrapper[4706]: I1127 07:33:40.705101 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-58swk_d987dc12-22dc-43cb-95dd-c40d07ca3786/kube-rbac-proxy/0.log" Nov 27 07:33:40 crc kubenswrapper[4706]: I1127 07:33:40.723408 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-58swk_d987dc12-22dc-43cb-95dd-c40d07ca3786/controller/0.log" Nov 27 07:33:40 crc kubenswrapper[4706]: I1127 07:33:40.828839 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.011641 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.022512 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.049549 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.060515 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.186424 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.201505 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.227485 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.246945 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.424807 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.454714 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.466375 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.493151 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/controller/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.632031 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/frr-metrics/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.632777 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/kube-rbac-proxy/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.690789 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/kube-rbac-proxy-frr/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.827055 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/reloader/0.log" Nov 27 07:33:41 crc kubenswrapper[4706]: I1127 07:33:41.939261 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-6lz22_b893961c-2ac3-4a53-86c3-ffc1529ac013/frr-k8s-webhook-server/0.log" Nov 27 07:33:42 crc kubenswrapper[4706]: I1127 07:33:42.092578 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79b86cb64-xd9k5_4941533e-f387-4b8b-96fd-933271e8f532/manager/0.log" Nov 27 07:33:42 crc kubenswrapper[4706]: I1127 07:33:42.210246 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/frr/0.log" Nov 27 07:33:42 crc kubenswrapper[4706]: I1127 07:33:42.218878 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-69f857b4f-nkg76_73781e8a-5a08-4e58-b974-2c9ef4242257/webhook-server/0.log" Nov 27 07:33:42 crc kubenswrapper[4706]: I1127 07:33:42.353656 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2r7r8_456beb89-e656-49ac-b6d2-ad92029ae549/kube-rbac-proxy/0.log" Nov 27 07:33:42 crc kubenswrapper[4706]: I1127 07:33:42.513134 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2r7r8_456beb89-e656-49ac-b6d2-ad92029ae549/speaker/0.log" Nov 27 07:33:42 crc kubenswrapper[4706]: I1127 07:33:42.779865 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:33:42 crc kubenswrapper[4706]: E1127 07:33:42.780475 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:33:46 crc kubenswrapper[4706]: E1127 07:33:46.601129 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:33:46 crc kubenswrapper[4706]: E1127 07:33:46.601566 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:34:18.601544972 +0000 UTC m=+1542.491135872 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:33:46 crc kubenswrapper[4706]: E1127 07:33:46.601163 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:33:46 crc kubenswrapper[4706]: E1127 07:33:46.602001 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:34:18.601989196 +0000 UTC m=+1542.491580006 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:33:53 crc kubenswrapper[4706]: I1127 07:33:53.565000 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_2482c5b8-e170-4a54-9505-d422bc31d390/openstackclient/0.log" Nov 27 07:33:56 crc kubenswrapper[4706]: I1127 07:33:56.781739 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:33:56 crc kubenswrapper[4706]: E1127 07:33:56.782681 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:34:04 crc kubenswrapper[4706]: I1127 07:34:04.800630 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/util/0.log" Nov 27 07:34:04 crc kubenswrapper[4706]: I1127 07:34:04.962460 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/util/0.log" Nov 27 07:34:04 crc kubenswrapper[4706]: I1127 07:34:04.981790 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/pull/0.log" Nov 27 07:34:04 crc kubenswrapper[4706]: I1127 07:34:04.984012 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/pull/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.177451 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/util/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.187086 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/pull/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.188542 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/extract/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.349580 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-utilities/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.497638 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-content/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.506333 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-utilities/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.526699 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-content/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.645655 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-utilities/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.665107 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-content/0.log" Nov 27 07:34:05 crc kubenswrapper[4706]: I1127 07:34:05.841851 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2t2d_20f404ea-f279-4dc8-9a56-0cf44343ede6/extract-utilities/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.028476 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/registry-server/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.057069 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2t2d_20f404ea-f279-4dc8-9a56-0cf44343ede6/extract-utilities/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.069689 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2t2d_20f404ea-f279-4dc8-9a56-0cf44343ede6/extract-content/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.101271 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2t2d_20f404ea-f279-4dc8-9a56-0cf44343ede6/extract-content/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.229719 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2t2d_20f404ea-f279-4dc8-9a56-0cf44343ede6/extract-content/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.253371 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2t2d_20f404ea-f279-4dc8-9a56-0cf44343ede6/extract-utilities/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.466306 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fzsrg_bd7353c5-075d-4f1d-b18e-2da68cc06652/marketplace-operator/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.560183 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2t2d_20f404ea-f279-4dc8-9a56-0cf44343ede6/registry-server/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.563367 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-utilities/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.655856 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-utilities/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.701813 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-content/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.719518 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-content/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.880440 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-content/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.896926 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-utilities/0.log" Nov 27 07:34:06 crc kubenswrapper[4706]: I1127 07:34:06.956585 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/registry-server/0.log" Nov 27 07:34:07 crc kubenswrapper[4706]: I1127 07:34:07.058074 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-utilities/0.log" Nov 27 07:34:07 crc kubenswrapper[4706]: I1127 07:34:07.201566 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-content/0.log" Nov 27 07:34:07 crc kubenswrapper[4706]: I1127 07:34:07.201627 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-content/0.log" Nov 27 07:34:07 crc kubenswrapper[4706]: I1127 07:34:07.221145 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-utilities/0.log" Nov 27 07:34:07 crc kubenswrapper[4706]: I1127 07:34:07.406407 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-content/0.log" Nov 27 07:34:07 crc kubenswrapper[4706]: I1127 07:34:07.416129 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-utilities/0.log" Nov 27 07:34:07 crc kubenswrapper[4706]: I1127 07:34:07.643861 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/registry-server/0.log" Nov 27 07:34:12 crc kubenswrapper[4706]: I1127 07:34:12.778141 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:34:12 crc kubenswrapper[4706]: E1127 07:34:12.779058 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.224089 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wtjlg"] Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.225537 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.259179 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtjlg"] Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.293935 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpb2k\" (UniqueName: \"kubernetes.io/projected/8de0e73e-fecb-4092-84b1-d977ca45c5cc-kube-api-access-fpb2k\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.294024 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-catalog-content\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.294364 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-utilities\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.396290 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpb2k\" (UniqueName: \"kubernetes.io/projected/8de0e73e-fecb-4092-84b1-d977ca45c5cc-kube-api-access-fpb2k\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.396413 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-catalog-content\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.396473 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-utilities\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.397055 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-catalog-content\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.397144 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-utilities\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.414359 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpb2k\" (UniqueName: \"kubernetes.io/projected/8de0e73e-fecb-4092-84b1-d977ca45c5cc-kube-api-access-fpb2k\") pod \"redhat-operators-wtjlg\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.543397 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:16 crc kubenswrapper[4706]: I1127 07:34:16.969169 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtjlg"] Nov 27 07:34:17 crc kubenswrapper[4706]: I1127 07:34:17.916882 4706 generic.go:334] "Generic (PLEG): container finished" podID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerID="759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd" exitCode=0 Nov 27 07:34:17 crc kubenswrapper[4706]: I1127 07:34:17.916930 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjlg" event={"ID":"8de0e73e-fecb-4092-84b1-d977ca45c5cc","Type":"ContainerDied","Data":"759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd"} Nov 27 07:34:17 crc kubenswrapper[4706]: I1127 07:34:17.917191 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjlg" event={"ID":"8de0e73e-fecb-4092-84b1-d977ca45c5cc","Type":"ContainerStarted","Data":"f20413a7e448c311fa545f3f26af331ff11fefb89d0d18635cfa1666b57bdd04"} Nov 27 07:34:18 crc kubenswrapper[4706]: E1127 07:34:18.629713 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:34:18 crc kubenswrapper[4706]: E1127 07:34:18.629784 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:35:22.62976933 +0000 UTC m=+1606.519360150 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:34:18 crc kubenswrapper[4706]: E1127 07:34:18.629711 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:34:18 crc kubenswrapper[4706]: E1127 07:34:18.629905 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:35:22.629885162 +0000 UTC m=+1606.519475982 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:34:19 crc kubenswrapper[4706]: I1127 07:34:19.930960 4706 generic.go:334] "Generic (PLEG): container finished" podID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerID="c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece" exitCode=0 Nov 27 07:34:19 crc kubenswrapper[4706]: I1127 07:34:19.931027 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjlg" event={"ID":"8de0e73e-fecb-4092-84b1-d977ca45c5cc","Type":"ContainerDied","Data":"c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece"} Nov 27 07:34:20 crc kubenswrapper[4706]: I1127 07:34:20.944638 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjlg" event={"ID":"8de0e73e-fecb-4092-84b1-d977ca45c5cc","Type":"ContainerStarted","Data":"01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a"} Nov 27 07:34:20 crc kubenswrapper[4706]: I1127 07:34:20.969250 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wtjlg" podStartSLOduration=2.513011633 podStartE2EDuration="4.96923664s" podCreationTimestamp="2025-11-27 07:34:16 +0000 UTC" firstStartedPulling="2025-11-27 07:34:17.918664646 +0000 UTC m=+1541.808255456" lastFinishedPulling="2025-11-27 07:34:20.374889633 +0000 UTC m=+1544.264480463" observedRunningTime="2025-11-27 07:34:20.965070609 +0000 UTC m=+1544.854661419" watchObservedRunningTime="2025-11-27 07:34:20.96923664 +0000 UTC m=+1544.858827450" Nov 27 07:34:26 crc kubenswrapper[4706]: I1127 07:34:26.544504 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:26 crc kubenswrapper[4706]: I1127 07:34:26.544878 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:26 crc kubenswrapper[4706]: I1127 07:34:26.597122 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:27 crc kubenswrapper[4706]: I1127 07:34:27.038374 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:27 crc kubenswrapper[4706]: I1127 07:34:27.080659 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wtjlg"] Nov 27 07:34:27 crc kubenswrapper[4706]: I1127 07:34:27.776252 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:34:27 crc kubenswrapper[4706]: E1127 07:34:27.776486 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:34:28 crc kubenswrapper[4706]: I1127 07:34:28.988988 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wtjlg" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="registry-server" containerID="cri-o://01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a" gracePeriod=2 Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.414004 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.472427 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-catalog-content\") pod \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.472578 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpb2k\" (UniqueName: \"kubernetes.io/projected/8de0e73e-fecb-4092-84b1-d977ca45c5cc-kube-api-access-fpb2k\") pod \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.472637 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-utilities\") pod \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\" (UID: \"8de0e73e-fecb-4092-84b1-d977ca45c5cc\") " Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.473994 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-utilities" (OuterVolumeSpecName: "utilities") pod "8de0e73e-fecb-4092-84b1-d977ca45c5cc" (UID: "8de0e73e-fecb-4092-84b1-d977ca45c5cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.494259 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8de0e73e-fecb-4092-84b1-d977ca45c5cc-kube-api-access-fpb2k" (OuterVolumeSpecName: "kube-api-access-fpb2k") pod "8de0e73e-fecb-4092-84b1-d977ca45c5cc" (UID: "8de0e73e-fecb-4092-84b1-d977ca45c5cc"). InnerVolumeSpecName "kube-api-access-fpb2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.573730 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpb2k\" (UniqueName: \"kubernetes.io/projected/8de0e73e-fecb-4092-84b1-d977ca45c5cc-kube-api-access-fpb2k\") on node \"crc\" DevicePath \"\"" Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.573758 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.998688 4706 generic.go:334] "Generic (PLEG): container finished" podID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerID="01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a" exitCode=0 Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.998889 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjlg" event={"ID":"8de0e73e-fecb-4092-84b1-d977ca45c5cc","Type":"ContainerDied","Data":"01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a"} Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.999210 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjlg" event={"ID":"8de0e73e-fecb-4092-84b1-d977ca45c5cc","Type":"ContainerDied","Data":"f20413a7e448c311fa545f3f26af331ff11fefb89d0d18635cfa1666b57bdd04"} Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.999251 4706 scope.go:117] "RemoveContainer" containerID="01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a" Nov 27 07:34:29 crc kubenswrapper[4706]: I1127 07:34:29.998986 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjlg" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.029868 4706 scope.go:117] "RemoveContainer" containerID="c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.073678 4706 scope.go:117] "RemoveContainer" containerID="759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.104934 4706 scope.go:117] "RemoveContainer" containerID="01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a" Nov 27 07:34:30 crc kubenswrapper[4706]: E1127 07:34:30.105634 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a\": container with ID starting with 01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a not found: ID does not exist" containerID="01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.105689 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a"} err="failed to get container status \"01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a\": rpc error: code = NotFound desc = could not find container \"01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a\": container with ID starting with 01c07f892deb1878cc64e694092f7b46ab854e637a799279f6a77c9113b1e01a not found: ID does not exist" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.105724 4706 scope.go:117] "RemoveContainer" containerID="c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece" Nov 27 07:34:30 crc kubenswrapper[4706]: E1127 07:34:30.109384 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece\": container with ID starting with c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece not found: ID does not exist" containerID="c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.109450 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece"} err="failed to get container status \"c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece\": rpc error: code = NotFound desc = could not find container \"c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece\": container with ID starting with c3ef257845943a45677d2b9383f64328760e14c73fc3ec04d49458c08aa2bece not found: ID does not exist" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.109482 4706 scope.go:117] "RemoveContainer" containerID="759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd" Nov 27 07:34:30 crc kubenswrapper[4706]: E1127 07:34:30.113650 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd\": container with ID starting with 759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd not found: ID does not exist" containerID="759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd" Nov 27 07:34:30 crc kubenswrapper[4706]: I1127 07:34:30.113701 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd"} err="failed to get container status \"759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd\": rpc error: code = NotFound desc = could not find container \"759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd\": container with ID starting with 759d46b28c37cf74204b82c50ada749ab13eb271365697781c9d9cc8868ac1dd not found: ID does not exist" Nov 27 07:34:31 crc kubenswrapper[4706]: I1127 07:34:31.180868 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8de0e73e-fecb-4092-84b1-d977ca45c5cc" (UID: "8de0e73e-fecb-4092-84b1-d977ca45c5cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:34:31 crc kubenswrapper[4706]: I1127 07:34:31.196728 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8de0e73e-fecb-4092-84b1-d977ca45c5cc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:34:31 crc kubenswrapper[4706]: I1127 07:34:31.244139 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wtjlg"] Nov 27 07:34:31 crc kubenswrapper[4706]: I1127 07:34:31.249870 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wtjlg"] Nov 27 07:34:32 crc kubenswrapper[4706]: I1127 07:34:32.787858 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" path="/var/lib/kubelet/pods/8de0e73e-fecb-4092-84b1-d977ca45c5cc/volumes" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.280236 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qd9pp"] Nov 27 07:34:38 crc kubenswrapper[4706]: E1127 07:34:38.280833 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="extract-utilities" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.280849 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="extract-utilities" Nov 27 07:34:38 crc kubenswrapper[4706]: E1127 07:34:38.280870 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="extract-content" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.280877 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="extract-content" Nov 27 07:34:38 crc kubenswrapper[4706]: E1127 07:34:38.280897 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="registry-server" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.280905 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="registry-server" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.281028 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de0e73e-fecb-4092-84b1-d977ca45c5cc" containerName="registry-server" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.281932 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.303426 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qd9pp"] Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.401997 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34610560-6b02-4f75-b595-2f3f0bef67e3-utilities\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.402077 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5z8c\" (UniqueName: \"kubernetes.io/projected/34610560-6b02-4f75-b595-2f3f0bef67e3-kube-api-access-x5z8c\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.402150 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34610560-6b02-4f75-b595-2f3f0bef67e3-catalog-content\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.503088 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34610560-6b02-4f75-b595-2f3f0bef67e3-catalog-content\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.503172 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34610560-6b02-4f75-b595-2f3f0bef67e3-utilities\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.503214 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5z8c\" (UniqueName: \"kubernetes.io/projected/34610560-6b02-4f75-b595-2f3f0bef67e3-kube-api-access-x5z8c\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.503691 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34610560-6b02-4f75-b595-2f3f0bef67e3-catalog-content\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.503786 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34610560-6b02-4f75-b595-2f3f0bef67e3-utilities\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.531972 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5z8c\" (UniqueName: \"kubernetes.io/projected/34610560-6b02-4f75-b595-2f3f0bef67e3-kube-api-access-x5z8c\") pod \"community-operators-qd9pp\" (UID: \"34610560-6b02-4f75-b595-2f3f0bef67e3\") " pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:38 crc kubenswrapper[4706]: I1127 07:34:38.603731 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:39 crc kubenswrapper[4706]: I1127 07:34:39.058391 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qd9pp"] Nov 27 07:34:39 crc kubenswrapper[4706]: I1127 07:34:39.653893 4706 scope.go:117] "RemoveContainer" containerID="0b877983289f5c31f7ded4cf7191cc057ee865bc7c2d08a08bc6f39b8eb3a394" Nov 27 07:34:39 crc kubenswrapper[4706]: I1127 07:34:39.675695 4706 scope.go:117] "RemoveContainer" containerID="e83fb3ceda05b3ec275792009786ee6c28efc49055f1a64d88d8b83e14459425" Nov 27 07:34:39 crc kubenswrapper[4706]: I1127 07:34:39.690790 4706 scope.go:117] "RemoveContainer" containerID="3d2a2366c863bf89a56cda712ace9b140dddaf7bc03f07ce8f28ff89074b529d" Nov 27 07:34:39 crc kubenswrapper[4706]: I1127 07:34:39.741868 4706 scope.go:117] "RemoveContainer" containerID="cfcf7879f8f1b32335b57fefdc9fefa2104a7bdf7df49cca449cbdced5505fac" Nov 27 07:34:40 crc kubenswrapper[4706]: I1127 07:34:40.082930 4706 generic.go:334] "Generic (PLEG): container finished" podID="34610560-6b02-4f75-b595-2f3f0bef67e3" containerID="b923e4dc09c97b9d9b0694673d0c90f06ffa350c30269265c7c9031d52efd8ab" exitCode=0 Nov 27 07:34:40 crc kubenswrapper[4706]: I1127 07:34:40.082988 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd9pp" event={"ID":"34610560-6b02-4f75-b595-2f3f0bef67e3","Type":"ContainerDied","Data":"b923e4dc09c97b9d9b0694673d0c90f06ffa350c30269265c7c9031d52efd8ab"} Nov 27 07:34:40 crc kubenswrapper[4706]: I1127 07:34:40.083018 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd9pp" event={"ID":"34610560-6b02-4f75-b595-2f3f0bef67e3","Type":"ContainerStarted","Data":"88f7734ffb91dcebcd25585a0f8eab9eaf017cfc9b8408f4079cd17a88e6ea2b"} Nov 27 07:34:41 crc kubenswrapper[4706]: I1127 07:34:41.777531 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:34:41 crc kubenswrapper[4706]: E1127 07:34:41.777773 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:34:44 crc kubenswrapper[4706]: I1127 07:34:44.106077 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd9pp" event={"ID":"34610560-6b02-4f75-b595-2f3f0bef67e3","Type":"ContainerStarted","Data":"e76c4549c98f6731d13390cbe6a9f5e87836bc50a9c4bafc766c734efdc57644"} Nov 27 07:34:45 crc kubenswrapper[4706]: I1127 07:34:45.114910 4706 generic.go:334] "Generic (PLEG): container finished" podID="34610560-6b02-4f75-b595-2f3f0bef67e3" containerID="e76c4549c98f6731d13390cbe6a9f5e87836bc50a9c4bafc766c734efdc57644" exitCode=0 Nov 27 07:34:45 crc kubenswrapper[4706]: I1127 07:34:45.115876 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd9pp" event={"ID":"34610560-6b02-4f75-b595-2f3f0bef67e3","Type":"ContainerDied","Data":"e76c4549c98f6731d13390cbe6a9f5e87836bc50a9c4bafc766c734efdc57644"} Nov 27 07:34:46 crc kubenswrapper[4706]: I1127 07:34:46.138111 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd9pp" event={"ID":"34610560-6b02-4f75-b595-2f3f0bef67e3","Type":"ContainerStarted","Data":"e3158f69721073f3074f5b789ebdf67d84e58be81c51cebf7b3bc989b43e0ec8"} Nov 27 07:34:46 crc kubenswrapper[4706]: I1127 07:34:46.167140 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qd9pp" podStartSLOduration=2.592535137 podStartE2EDuration="8.167119964s" podCreationTimestamp="2025-11-27 07:34:38 +0000 UTC" firstStartedPulling="2025-11-27 07:34:40.084475996 +0000 UTC m=+1563.974066806" lastFinishedPulling="2025-11-27 07:34:45.659060813 +0000 UTC m=+1569.548651633" observedRunningTime="2025-11-27 07:34:46.160868617 +0000 UTC m=+1570.050459477" watchObservedRunningTime="2025-11-27 07:34:46.167119964 +0000 UTC m=+1570.056710784" Nov 27 07:34:48 crc kubenswrapper[4706]: I1127 07:34:48.603867 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:48 crc kubenswrapper[4706]: I1127 07:34:48.604195 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:48 crc kubenswrapper[4706]: I1127 07:34:48.670109 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.648928 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xzltz"] Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.650141 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.670483 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzltz"] Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.777590 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:34:52 crc kubenswrapper[4706]: E1127 07:34:52.778274 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.842272 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-catalog-content\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.842832 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-utilities\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.842928 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzvzb\" (UniqueName: \"kubernetes.io/projected/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-kube-api-access-dzvzb\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.944073 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-catalog-content\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.944124 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-utilities\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.944152 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzvzb\" (UniqueName: \"kubernetes.io/projected/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-kube-api-access-dzvzb\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.944574 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-utilities\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.944656 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-catalog-content\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:52 crc kubenswrapper[4706]: I1127 07:34:52.964058 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzvzb\" (UniqueName: \"kubernetes.io/projected/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-kube-api-access-dzvzb\") pod \"certified-operators-xzltz\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:53 crc kubenswrapper[4706]: I1127 07:34:53.014440 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:34:53 crc kubenswrapper[4706]: I1127 07:34:53.261226 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzltz"] Nov 27 07:34:54 crc kubenswrapper[4706]: I1127 07:34:54.197833 4706 generic.go:334] "Generic (PLEG): container finished" podID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerID="b80e0852a7fff52d1cd86a93d7d7e576c7a3edb068c90b60a543a012aae838ac" exitCode=0 Nov 27 07:34:54 crc kubenswrapper[4706]: I1127 07:34:54.197909 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzltz" event={"ID":"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02","Type":"ContainerDied","Data":"b80e0852a7fff52d1cd86a93d7d7e576c7a3edb068c90b60a543a012aae838ac"} Nov 27 07:34:54 crc kubenswrapper[4706]: I1127 07:34:54.197939 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzltz" event={"ID":"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02","Type":"ContainerStarted","Data":"2b529b578885e599bd7c3c1367c6ff8398abe4895e51f89479e9973072f17a27"} Nov 27 07:34:56 crc kubenswrapper[4706]: I1127 07:34:56.214268 4706 generic.go:334] "Generic (PLEG): container finished" podID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerID="5a1c939e2377fae10a4905456040228ea9450dedcc8eba4f2aa0850bed1e3c57" exitCode=0 Nov 27 07:34:56 crc kubenswrapper[4706]: I1127 07:34:56.214795 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzltz" event={"ID":"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02","Type":"ContainerDied","Data":"5a1c939e2377fae10a4905456040228ea9450dedcc8eba4f2aa0850bed1e3c57"} Nov 27 07:34:57 crc kubenswrapper[4706]: I1127 07:34:57.223038 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzltz" event={"ID":"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02","Type":"ContainerStarted","Data":"8a39e365a4f5cacee0b627ef0fe67e56d2c047d5157a385785bee18728a1b1f0"} Nov 27 07:34:57 crc kubenswrapper[4706]: I1127 07:34:57.240673 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xzltz" podStartSLOduration=2.775249391 podStartE2EDuration="5.240654695s" podCreationTimestamp="2025-11-27 07:34:52 +0000 UTC" firstStartedPulling="2025-11-27 07:34:54.199854998 +0000 UTC m=+1578.089445808" lastFinishedPulling="2025-11-27 07:34:56.665260302 +0000 UTC m=+1580.554851112" observedRunningTime="2025-11-27 07:34:57.239827969 +0000 UTC m=+1581.129418779" watchObservedRunningTime="2025-11-27 07:34:57.240654695 +0000 UTC m=+1581.130245505" Nov 27 07:34:58 crc kubenswrapper[4706]: I1127 07:34:58.660122 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qd9pp" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.246925 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6js"] Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.248724 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.255743 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6js"] Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.445061 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-catalog-content\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.445181 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-utilities\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.445213 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svnpr\" (UniqueName: \"kubernetes.io/projected/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-kube-api-access-svnpr\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.547214 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-catalog-content\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.547352 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-utilities\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.547389 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svnpr\" (UniqueName: \"kubernetes.io/projected/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-kube-api-access-svnpr\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.547903 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-catalog-content\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.547953 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-utilities\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.566927 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svnpr\" (UniqueName: \"kubernetes.io/projected/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-kube-api-access-svnpr\") pod \"redhat-marketplace-pd6js\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.579796 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:34:59 crc kubenswrapper[4706]: I1127 07:34:59.781705 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6js"] Nov 27 07:35:00 crc kubenswrapper[4706]: E1127 07:35:00.061252 4706 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8df7cf9d_37c4_46a4_b3b8_ea5ff602b5a9.slice/crio-conmon-3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8df7cf9d_37c4_46a4_b3b8_ea5ff602b5a9.slice/crio-3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4.scope\": RecentStats: unable to find data in memory cache]" Nov 27 07:35:00 crc kubenswrapper[4706]: I1127 07:35:00.255911 4706 generic.go:334] "Generic (PLEG): container finished" podID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerID="3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4" exitCode=0 Nov 27 07:35:00 crc kubenswrapper[4706]: I1127 07:35:00.255980 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6js" event={"ID":"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9","Type":"ContainerDied","Data":"3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4"} Nov 27 07:35:00 crc kubenswrapper[4706]: I1127 07:35:00.256009 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6js" event={"ID":"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9","Type":"ContainerStarted","Data":"8b489c0fe4de7a0a002b88e054e95922ae68bd3771f12bdac87fbbdca205dd48"} Nov 27 07:35:00 crc kubenswrapper[4706]: I1127 07:35:00.663691 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qd9pp"] Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.028190 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d2t2d"] Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.028876 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d2t2d" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="registry-server" containerID="cri-o://6c694b87d3e240f20100347fc67d101de4741cab8a4788abc2d431559a212d51" gracePeriod=2 Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.268331 4706 generic.go:334] "Generic (PLEG): container finished" podID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerID="cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82" exitCode=0 Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.268418 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6js" event={"ID":"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9","Type":"ContainerDied","Data":"cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82"} Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.278793 4706 generic.go:334] "Generic (PLEG): container finished" podID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerID="6c694b87d3e240f20100347fc67d101de4741cab8a4788abc2d431559a212d51" exitCode=0 Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.278846 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2t2d" event={"ID":"20f404ea-f279-4dc8-9a56-0cf44343ede6","Type":"ContainerDied","Data":"6c694b87d3e240f20100347fc67d101de4741cab8a4788abc2d431559a212d51"} Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.412535 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.472803 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-utilities\") pod \"20f404ea-f279-4dc8-9a56-0cf44343ede6\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.472887 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-catalog-content\") pod \"20f404ea-f279-4dc8-9a56-0cf44343ede6\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.472984 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw9xz\" (UniqueName: \"kubernetes.io/projected/20f404ea-f279-4dc8-9a56-0cf44343ede6-kube-api-access-qw9xz\") pod \"20f404ea-f279-4dc8-9a56-0cf44343ede6\" (UID: \"20f404ea-f279-4dc8-9a56-0cf44343ede6\") " Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.473564 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-utilities" (OuterVolumeSpecName: "utilities") pod "20f404ea-f279-4dc8-9a56-0cf44343ede6" (UID: "20f404ea-f279-4dc8-9a56-0cf44343ede6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.493472 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f404ea-f279-4dc8-9a56-0cf44343ede6-kube-api-access-qw9xz" (OuterVolumeSpecName: "kube-api-access-qw9xz") pod "20f404ea-f279-4dc8-9a56-0cf44343ede6" (UID: "20f404ea-f279-4dc8-9a56-0cf44343ede6"). InnerVolumeSpecName "kube-api-access-qw9xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.526621 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20f404ea-f279-4dc8-9a56-0cf44343ede6" (UID: "20f404ea-f279-4dc8-9a56-0cf44343ede6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.574959 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw9xz\" (UniqueName: \"kubernetes.io/projected/20f404ea-f279-4dc8-9a56-0cf44343ede6-kube-api-access-qw9xz\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.575009 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:01 crc kubenswrapper[4706]: I1127 07:35:01.575022 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20f404ea-f279-4dc8-9a56-0cf44343ede6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.286496 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6js" event={"ID":"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9","Type":"ContainerStarted","Data":"8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce"} Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.288753 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2t2d" event={"ID":"20f404ea-f279-4dc8-9a56-0cf44343ede6","Type":"ContainerDied","Data":"bbf35eedbf9f2a1d026d922aa2566ece573c8ccb850804accce4963be65bd981"} Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.288800 4706 scope.go:117] "RemoveContainer" containerID="6c694b87d3e240f20100347fc67d101de4741cab8a4788abc2d431559a212d51" Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.288825 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2t2d" Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.320159 4706 scope.go:117] "RemoveContainer" containerID="f8acfae962b6171d0184004287668879c8c0e9d26a2d6a6d1a417335912c65bd" Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.334332 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pd6js" podStartSLOduration=1.7587204829999998 podStartE2EDuration="3.334312656s" podCreationTimestamp="2025-11-27 07:34:59 +0000 UTC" firstStartedPulling="2025-11-27 07:35:00.257783817 +0000 UTC m=+1584.147374627" lastFinishedPulling="2025-11-27 07:35:01.833376 +0000 UTC m=+1585.722966800" observedRunningTime="2025-11-27 07:35:02.318769197 +0000 UTC m=+1586.208360017" watchObservedRunningTime="2025-11-27 07:35:02.334312656 +0000 UTC m=+1586.223903486" Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.335388 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d2t2d"] Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.342918 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d2t2d"] Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.355394 4706 scope.go:117] "RemoveContainer" containerID="57e4a22e7aa094044541569800f8c62a91c13c805e60f955a7db0fbc68a17660" Nov 27 07:35:02 crc kubenswrapper[4706]: I1127 07:35:02.787516 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" path="/var/lib/kubelet/pods/20f404ea-f279-4dc8-9a56-0cf44343ede6/volumes" Nov 27 07:35:03 crc kubenswrapper[4706]: I1127 07:35:03.015302 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:35:03 crc kubenswrapper[4706]: I1127 07:35:03.015381 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:35:03 crc kubenswrapper[4706]: I1127 07:35:03.078709 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:35:03 crc kubenswrapper[4706]: I1127 07:35:03.331883 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:35:05 crc kubenswrapper[4706]: I1127 07:35:05.635593 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzltz"] Nov 27 07:35:05 crc kubenswrapper[4706]: I1127 07:35:05.636060 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xzltz" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="registry-server" containerID="cri-o://8a39e365a4f5cacee0b627ef0fe67e56d2c047d5157a385785bee18728a1b1f0" gracePeriod=2 Nov 27 07:35:06 crc kubenswrapper[4706]: I1127 07:35:06.780983 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:35:06 crc kubenswrapper[4706]: E1127 07:35:06.781526 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.334332 4706 generic.go:334] "Generic (PLEG): container finished" podID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerID="8a39e365a4f5cacee0b627ef0fe67e56d2c047d5157a385785bee18728a1b1f0" exitCode=0 Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.334464 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzltz" event={"ID":"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02","Type":"ContainerDied","Data":"8a39e365a4f5cacee0b627ef0fe67e56d2c047d5157a385785bee18728a1b1f0"} Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.334742 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzltz" event={"ID":"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02","Type":"ContainerDied","Data":"2b529b578885e599bd7c3c1367c6ff8398abe4895e51f89479e9973072f17a27"} Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.334768 4706 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b529b578885e599bd7c3c1367c6ff8398abe4895e51f89479e9973072f17a27" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.363173 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.464137 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-utilities\") pod \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.464412 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-catalog-content\") pod \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.464481 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzvzb\" (UniqueName: \"kubernetes.io/projected/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-kube-api-access-dzvzb\") pod \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\" (UID: \"43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02\") " Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.465199 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-utilities" (OuterVolumeSpecName: "utilities") pod "43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" (UID: "43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.484596 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-kube-api-access-dzvzb" (OuterVolumeSpecName: "kube-api-access-dzvzb") pod "43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" (UID: "43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02"). InnerVolumeSpecName "kube-api-access-dzvzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.518339 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" (UID: "43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.565568 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.565597 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:07 crc kubenswrapper[4706]: I1127 07:35:07.565610 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzvzb\" (UniqueName: \"kubernetes.io/projected/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02-kube-api-access-dzvzb\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:08 crc kubenswrapper[4706]: I1127 07:35:08.341278 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzltz" Nov 27 07:35:08 crc kubenswrapper[4706]: I1127 07:35:08.386045 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzltz"] Nov 27 07:35:08 crc kubenswrapper[4706]: I1127 07:35:08.392398 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xzltz"] Nov 27 07:35:08 crc kubenswrapper[4706]: I1127 07:35:08.787513 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" path="/var/lib/kubelet/pods/43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02/volumes" Nov 27 07:35:09 crc kubenswrapper[4706]: I1127 07:35:09.581150 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:35:09 crc kubenswrapper[4706]: I1127 07:35:09.581854 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:35:09 crc kubenswrapper[4706]: I1127 07:35:09.640968 4706 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:35:10 crc kubenswrapper[4706]: I1127 07:35:10.410822 4706 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:35:11 crc kubenswrapper[4706]: I1127 07:35:11.365158 4706 generic.go:334] "Generic (PLEG): container finished" podID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerID="8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603" exitCode=0 Nov 27 07:35:11 crc kubenswrapper[4706]: I1127 07:35:11.365327 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jrcsw/must-gather-d65pf" event={"ID":"abf4bcc1-8309-4b72-acd7-7b246a41eff6","Type":"ContainerDied","Data":"8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603"} Nov 27 07:35:11 crc kubenswrapper[4706]: I1127 07:35:11.366608 4706 scope.go:117] "RemoveContainer" containerID="8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603" Nov 27 07:35:11 crc kubenswrapper[4706]: I1127 07:35:11.629853 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6js"] Nov 27 07:35:12 crc kubenswrapper[4706]: I1127 07:35:12.311856 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jrcsw_must-gather-d65pf_abf4bcc1-8309-4b72-acd7-7b246a41eff6/gather/0.log" Nov 27 07:35:12 crc kubenswrapper[4706]: I1127 07:35:12.372846 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pd6js" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="registry-server" containerID="cri-o://8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce" gracePeriod=2 Nov 27 07:35:12 crc kubenswrapper[4706]: I1127 07:35:12.885389 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:35:12 crc kubenswrapper[4706]: I1127 07:35:12.952450 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-catalog-content\") pod \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " Nov 27 07:35:12 crc kubenswrapper[4706]: I1127 07:35:12.952490 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svnpr\" (UniqueName: \"kubernetes.io/projected/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-kube-api-access-svnpr\") pod \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " Nov 27 07:35:12 crc kubenswrapper[4706]: I1127 07:35:12.957783 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-kube-api-access-svnpr" (OuterVolumeSpecName: "kube-api-access-svnpr") pod "8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" (UID: "8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9"). InnerVolumeSpecName "kube-api-access-svnpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:35:12 crc kubenswrapper[4706]: I1127 07:35:12.968212 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" (UID: "8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.053352 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-utilities\") pod \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\" (UID: \"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9\") " Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.053618 4706 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.053634 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svnpr\" (UniqueName: \"kubernetes.io/projected/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-kube-api-access-svnpr\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.054166 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-utilities" (OuterVolumeSpecName: "utilities") pod "8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" (UID: "8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.155613 4706 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.383312 4706 generic.go:334] "Generic (PLEG): container finished" podID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerID="8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce" exitCode=0 Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.383367 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6js" event={"ID":"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9","Type":"ContainerDied","Data":"8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce"} Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.383389 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pd6js" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.383404 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pd6js" event={"ID":"8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9","Type":"ContainerDied","Data":"8b489c0fe4de7a0a002b88e054e95922ae68bd3771f12bdac87fbbdca205dd48"} Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.383436 4706 scope.go:117] "RemoveContainer" containerID="8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.404980 4706 scope.go:117] "RemoveContainer" containerID="cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.429944 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6js"] Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.435421 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pd6js"] Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.444428 4706 scope.go:117] "RemoveContainer" containerID="3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.468544 4706 scope.go:117] "RemoveContainer" containerID="8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce" Nov 27 07:35:13 crc kubenswrapper[4706]: E1127 07:35:13.469360 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce\": container with ID starting with 8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce not found: ID does not exist" containerID="8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.469410 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce"} err="failed to get container status \"8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce\": rpc error: code = NotFound desc = could not find container \"8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce\": container with ID starting with 8e809955aafe530d094beabb7485558e86e077a34a878d22cb762c9f5b0ad4ce not found: ID does not exist" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.469469 4706 scope.go:117] "RemoveContainer" containerID="cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82" Nov 27 07:35:13 crc kubenswrapper[4706]: E1127 07:35:13.470080 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82\": container with ID starting with cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82 not found: ID does not exist" containerID="cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.470125 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82"} err="failed to get container status \"cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82\": rpc error: code = NotFound desc = could not find container \"cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82\": container with ID starting with cc9b2b0634d54f365607e5867d410de4fe06affc8ece1c0cdc3107f20ddb6e82 not found: ID does not exist" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.470157 4706 scope.go:117] "RemoveContainer" containerID="3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4" Nov 27 07:35:13 crc kubenswrapper[4706]: E1127 07:35:13.470775 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4\": container with ID starting with 3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4 not found: ID does not exist" containerID="3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4" Nov 27 07:35:13 crc kubenswrapper[4706]: I1127 07:35:13.471534 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4"} err="failed to get container status \"3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4\": rpc error: code = NotFound desc = could not find container \"3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4\": container with ID starting with 3e59675ca7ae2b40a03fdc9bd532b0c48fe9ffac340cc442c62f3c1616f955f4 not found: ID does not exist" Nov 27 07:35:14 crc kubenswrapper[4706]: I1127 07:35:14.785852 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" path="/var/lib/kubelet/pods/8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9/volumes" Nov 27 07:35:17 crc kubenswrapper[4706]: I1127 07:35:17.776921 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:35:17 crc kubenswrapper[4706]: E1127 07:35:17.777749 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.463020 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jrcsw/must-gather-d65pf"] Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.463546 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jrcsw/must-gather-d65pf" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerName="copy" containerID="cri-o://c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d" gracePeriod=2 Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.467797 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jrcsw/must-gather-d65pf"] Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.899759 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jrcsw_must-gather-d65pf_abf4bcc1-8309-4b72-acd7-7b246a41eff6/copy/0.log" Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.900332 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.905030 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/abf4bcc1-8309-4b72-acd7-7b246a41eff6-must-gather-output\") pod \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.905104 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxmdh\" (UniqueName: \"kubernetes.io/projected/abf4bcc1-8309-4b72-acd7-7b246a41eff6-kube-api-access-mxmdh\") pod \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\" (UID: \"abf4bcc1-8309-4b72-acd7-7b246a41eff6\") " Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.910324 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf4bcc1-8309-4b72-acd7-7b246a41eff6-kube-api-access-mxmdh" (OuterVolumeSpecName: "kube-api-access-mxmdh") pod "abf4bcc1-8309-4b72-acd7-7b246a41eff6" (UID: "abf4bcc1-8309-4b72-acd7-7b246a41eff6"). InnerVolumeSpecName "kube-api-access-mxmdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:35:19 crc kubenswrapper[4706]: I1127 07:35:19.973142 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf4bcc1-8309-4b72-acd7-7b246a41eff6-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "abf4bcc1-8309-4b72-acd7-7b246a41eff6" (UID: "abf4bcc1-8309-4b72-acd7-7b246a41eff6"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.006521 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxmdh\" (UniqueName: \"kubernetes.io/projected/abf4bcc1-8309-4b72-acd7-7b246a41eff6-kube-api-access-mxmdh\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.006768 4706 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/abf4bcc1-8309-4b72-acd7-7b246a41eff6-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.445426 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jrcsw_must-gather-d65pf_abf4bcc1-8309-4b72-acd7-7b246a41eff6/copy/0.log" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.446640 4706 generic.go:334] "Generic (PLEG): container finished" podID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerID="c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d" exitCode=143 Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.446752 4706 scope.go:117] "RemoveContainer" containerID="c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.446827 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jrcsw/must-gather-d65pf" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.466057 4706 scope.go:117] "RemoveContainer" containerID="8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.522116 4706 scope.go:117] "RemoveContainer" containerID="c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d" Nov 27 07:35:20 crc kubenswrapper[4706]: E1127 07:35:20.522776 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d\": container with ID starting with c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d not found: ID does not exist" containerID="c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.522818 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d"} err="failed to get container status \"c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d\": rpc error: code = NotFound desc = could not find container \"c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d\": container with ID starting with c1a0e7fae90b4a985fcef2336e48e05dac58299b1d464160a11df3b17441f40d not found: ID does not exist" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.522845 4706 scope.go:117] "RemoveContainer" containerID="8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603" Nov 27 07:35:20 crc kubenswrapper[4706]: E1127 07:35:20.523190 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603\": container with ID starting with 8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603 not found: ID does not exist" containerID="8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.523386 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603"} err="failed to get container status \"8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603\": rpc error: code = NotFound desc = could not find container \"8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603\": container with ID starting with 8bdcbdd091f810f7a2f35abef98cc226b1da4ad80c46944c051c1f021d270603 not found: ID does not exist" Nov 27 07:35:20 crc kubenswrapper[4706]: I1127 07:35:20.785348 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" path="/var/lib/kubelet/pods/abf4bcc1-8309-4b72-acd7-7b246a41eff6/volumes" Nov 27 07:35:22 crc kubenswrapper[4706]: E1127 07:35:22.641589 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:35:22 crc kubenswrapper[4706]: E1127 07:35:22.641590 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:35:22 crc kubenswrapper[4706]: E1127 07:35:22.642023 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:37:24.641995563 +0000 UTC m=+1728.531586403 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:35:22 crc kubenswrapper[4706]: E1127 07:35:22.642124 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:37:24.642095885 +0000 UTC m=+1728.531686735 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:35:32 crc kubenswrapper[4706]: I1127 07:35:32.776578 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:35:32 crc kubenswrapper[4706]: E1127 07:35:32.777431 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:35:39 crc kubenswrapper[4706]: I1127 07:35:39.846686 4706 scope.go:117] "RemoveContainer" containerID="0a759fba5b0de9a2e4b623f0c831b18f663e33d44f411bb224458f47e7d934db" Nov 27 07:35:39 crc kubenswrapper[4706]: I1127 07:35:39.875572 4706 scope.go:117] "RemoveContainer" containerID="022b03133d72ddf87c7fa052b05bc029e5613859e2a9efcc37d46fb41348ae96" Nov 27 07:35:39 crc kubenswrapper[4706]: I1127 07:35:39.900681 4706 scope.go:117] "RemoveContainer" containerID="360bf9d54cb5135b9a7ee75e00f1e0790e98428f2a4162427097a7f65f4a09c4" Nov 27 07:35:39 crc kubenswrapper[4706]: I1127 07:35:39.946348 4706 scope.go:117] "RemoveContainer" containerID="a819058332255933355b0cc51c35255f77ceca9b20ebb06695156854d9aa72b6" Nov 27 07:35:44 crc kubenswrapper[4706]: I1127 07:35:44.777356 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:35:44 crc kubenswrapper[4706]: E1127 07:35:44.777733 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:35:59 crc kubenswrapper[4706]: I1127 07:35:59.777530 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:35:59 crc kubenswrapper[4706]: E1127 07:35:59.778805 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:36:14 crc kubenswrapper[4706]: I1127 07:36:14.777093 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:36:14 crc kubenswrapper[4706]: E1127 07:36:14.777687 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:36:29 crc kubenswrapper[4706]: I1127 07:36:29.777146 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:36:29 crc kubenswrapper[4706]: E1127 07:36:29.778187 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:36:40 crc kubenswrapper[4706]: I1127 07:36:40.050229 4706 scope.go:117] "RemoveContainer" containerID="96b638f010819e9ff57f884e5715c34289491bbb2242e29c18262152b06f7621" Nov 27 07:36:40 crc kubenswrapper[4706]: I1127 07:36:40.107452 4706 scope.go:117] "RemoveContainer" containerID="bd810a5801a6e24867ccfb8c4a0932f5ae2edea4a3acb69e4922725d53f1114d" Nov 27 07:36:40 crc kubenswrapper[4706]: I1127 07:36:40.128044 4706 scope.go:117] "RemoveContainer" containerID="f9e493bc129e15e013be1ae53108b37b262a81e065401f360568a1d57fc111a1" Nov 27 07:36:40 crc kubenswrapper[4706]: I1127 07:36:40.170045 4706 scope.go:117] "RemoveContainer" containerID="e486246d7f6dc59b55b26a87ad194529362f0f9e42095eda2b81f5ee782873ad" Nov 27 07:36:40 crc kubenswrapper[4706]: I1127 07:36:40.217099 4706 scope.go:117] "RemoveContainer" containerID="1e657a803e73c8309a59300617776cc8169e2d9c54cf6150e4e21e636ef8156c" Nov 27 07:36:40 crc kubenswrapper[4706]: I1127 07:36:40.240454 4706 scope.go:117] "RemoveContainer" containerID="da2bef8fec31533ea8254a7be675d2b633117bd57b9252efe993948959dd22e8" Nov 27 07:36:40 crc kubenswrapper[4706]: I1127 07:36:40.277368 4706 scope.go:117] "RemoveContainer" containerID="900f66aebd0ca0ecf88ac5851201956fb93ca9d8a5fd8e9f071f080eb3ce0876" Nov 27 07:36:42 crc kubenswrapper[4706]: I1127 07:36:42.777229 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:36:42 crc kubenswrapper[4706]: E1127 07:36:42.778013 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:36:56 crc kubenswrapper[4706]: I1127 07:36:56.781066 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:36:56 crc kubenswrapper[4706]: E1127 07:36:56.781906 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:37:10 crc kubenswrapper[4706]: I1127 07:37:10.777279 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:37:10 crc kubenswrapper[4706]: E1127 07:37:10.777992 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:37:23 crc kubenswrapper[4706]: I1127 07:37:23.776697 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:37:23 crc kubenswrapper[4706]: E1127 07:37:23.777410 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:37:24 crc kubenswrapper[4706]: E1127 07:37:24.658154 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:37:24 crc kubenswrapper[4706]: E1127 07:37:24.658161 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:37:24 crc kubenswrapper[4706]: E1127 07:37:24.658258 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:39:26.658238936 +0000 UTC m=+1850.547829756 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:37:24 crc kubenswrapper[4706]: E1127 07:37:24.658327 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:39:26.658306808 +0000 UTC m=+1850.547897648 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.880243 4706 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mg2kk/must-gather-hwfz2"] Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881395 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="extract-content" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881416 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="extract-content" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881439 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="extract-utilities" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881450 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="extract-utilities" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881466 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="extract-content" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881477 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="extract-content" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881494 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="extract-utilities" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881504 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="extract-utilities" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881525 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerName="gather" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881535 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerName="gather" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881549 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881559 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881575 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881586 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881619 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="extract-utilities" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881630 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="extract-utilities" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881659 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerName="copy" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881669 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerName="copy" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881684 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881695 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: E1127 07:37:33.881714 4706 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="extract-content" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881726 4706 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="extract-content" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881885 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f404ea-f279-4dc8-9a56-0cf44343ede6" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881906 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="8df7cf9d-37c4-46a4-b3b8-ea5ff602b5a9" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881924 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerName="gather" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881936 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf4bcc1-8309-4b72-acd7-7b246a41eff6" containerName="copy" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.881950 4706 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e2fbe0-0525-4ccc-8577-3f8e6b8a7f02" containerName="registry-server" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.883062 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.885053 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mg2kk"/"openshift-service-ca.crt" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.885720 4706 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mg2kk"/"kube-root-ca.crt" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.900285 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkwnt\" (UniqueName: \"kubernetes.io/projected/0835baf9-d98f-44a6-af44-32e172968f2d-kube-api-access-nkwnt\") pod \"must-gather-hwfz2\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.900817 4706 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0835baf9-d98f-44a6-af44-32e172968f2d-must-gather-output\") pod \"must-gather-hwfz2\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:33 crc kubenswrapper[4706]: I1127 07:37:33.910566 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mg2kk/must-gather-hwfz2"] Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.002015 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0835baf9-d98f-44a6-af44-32e172968f2d-must-gather-output\") pod \"must-gather-hwfz2\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.002063 4706 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkwnt\" (UniqueName: \"kubernetes.io/projected/0835baf9-d98f-44a6-af44-32e172968f2d-kube-api-access-nkwnt\") pod \"must-gather-hwfz2\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.002539 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0835baf9-d98f-44a6-af44-32e172968f2d-must-gather-output\") pod \"must-gather-hwfz2\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.025868 4706 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkwnt\" (UniqueName: \"kubernetes.io/projected/0835baf9-d98f-44a6-af44-32e172968f2d-kube-api-access-nkwnt\") pod \"must-gather-hwfz2\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.200647 4706 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.408638 4706 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mg2kk/must-gather-hwfz2"] Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.596713 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" event={"ID":"0835baf9-d98f-44a6-af44-32e172968f2d","Type":"ContainerStarted","Data":"9fc74699f95ef976d9db04dade030c1371907b20d3a05bdccff4104d064654db"} Nov 27 07:37:34 crc kubenswrapper[4706]: I1127 07:37:34.777514 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:37:34 crc kubenswrapper[4706]: E1127 07:37:34.777797 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:37:35 crc kubenswrapper[4706]: I1127 07:37:35.606095 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" event={"ID":"0835baf9-d98f-44a6-af44-32e172968f2d","Type":"ContainerStarted","Data":"0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156"} Nov 27 07:37:35 crc kubenswrapper[4706]: I1127 07:37:35.607663 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" event={"ID":"0835baf9-d98f-44a6-af44-32e172968f2d","Type":"ContainerStarted","Data":"86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc"} Nov 27 07:37:35 crc kubenswrapper[4706]: I1127 07:37:35.618960 4706 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" podStartSLOduration=2.618937852 podStartE2EDuration="2.618937852s" podCreationTimestamp="2025-11-27 07:37:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 07:37:35.617986073 +0000 UTC m=+1739.507576913" watchObservedRunningTime="2025-11-27 07:37:35.618937852 +0000 UTC m=+1739.508528672" Nov 27 07:37:49 crc kubenswrapper[4706]: I1127 07:37:49.776803 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:37:49 crc kubenswrapper[4706]: E1127 07:37:49.777664 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:38:00 crc kubenswrapper[4706]: I1127 07:38:00.776557 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:38:00 crc kubenswrapper[4706]: E1127 07:38:00.777334 4706 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c44hv_openshift-machine-config-operator(f89d12ee-9aae-48da-a7a6-130c7186f506)\"" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.207924 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/util/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.339057 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/util/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.395445 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/pull/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.395664 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/pull/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.494377 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/pull/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.529206 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/util/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.569847 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cqfrm6_82d9ba46-8fc0-462c-bc77-056ef3252965/extract/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.668325 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54fdcb5fdb-dn97q_0961e9eb-dadf-4e6a-bbbb-34cd06d1b683/manager/0.log" Nov 27 07:38:05 crc kubenswrapper[4706]: I1127 07:38:05.802329 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-jnvft_7e5fd17c-d97a-4826-bd68-cd8480f1363d/registry-server/0.log" Nov 27 07:38:15 crc kubenswrapper[4706]: I1127 07:38:15.776858 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:38:16 crc kubenswrapper[4706]: I1127 07:38:16.839200 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"df4cd0396271a310394a11f3f8671785c69118c72eb0d461caf52dc2d37d6433"} Nov 27 07:38:16 crc kubenswrapper[4706]: I1127 07:38:16.885458 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-c878d_dd986c7f-e3ae-4d24-8815-c38d546168e7/control-plane-machine-set-operator/0.log" Nov 27 07:38:17 crc kubenswrapper[4706]: I1127 07:38:17.006855 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r6h9j_0b4e82db-e890-4dc6-a6f0-441d78878459/kube-rbac-proxy/0.log" Nov 27 07:38:17 crc kubenswrapper[4706]: I1127 07:38:17.033846 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r6h9j_0b4e82db-e890-4dc6-a6f0-441d78878459/machine-api-operator/0.log" Nov 27 07:38:31 crc kubenswrapper[4706]: I1127 07:38:31.624305 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-58swk_d987dc12-22dc-43cb-95dd-c40d07ca3786/kube-rbac-proxy/0.log" Nov 27 07:38:31 crc kubenswrapper[4706]: I1127 07:38:31.663732 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-58swk_d987dc12-22dc-43cb-95dd-c40d07ca3786/controller/0.log" Nov 27 07:38:31 crc kubenswrapper[4706]: I1127 07:38:31.784141 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:38:31 crc kubenswrapper[4706]: I1127 07:38:31.993739 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.009825 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.150822 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.167175 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.327058 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.327582 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.328089 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.358116 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.489074 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-frr-files/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.489287 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-metrics/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.496125 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/cp-reloader/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.522351 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/controller/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.674301 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/frr-metrics/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.679412 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/kube-rbac-proxy/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.727783 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/kube-rbac-proxy-frr/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.909445 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-6lz22_b893961c-2ac3-4a53-86c3-ffc1529ac013/frr-k8s-webhook-server/0.log" Nov 27 07:38:32 crc kubenswrapper[4706]: I1127 07:38:32.924357 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/reloader/0.log" Nov 27 07:38:33 crc kubenswrapper[4706]: I1127 07:38:33.124753 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79b86cb64-xd9k5_4941533e-f387-4b8b-96fd-933271e8f532/manager/0.log" Nov 27 07:38:33 crc kubenswrapper[4706]: I1127 07:38:33.258279 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-69f857b4f-nkg76_73781e8a-5a08-4e58-b974-2c9ef4242257/webhook-server/0.log" Nov 27 07:38:33 crc kubenswrapper[4706]: I1127 07:38:33.297048 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2r7r8_456beb89-e656-49ac-b6d2-ad92029ae549/kube-rbac-proxy/0.log" Nov 27 07:38:33 crc kubenswrapper[4706]: I1127 07:38:33.401550 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bkj8q_de4c78be-fea0-452b-bfb7-59b423f0cdf7/frr/0.log" Nov 27 07:38:33 crc kubenswrapper[4706]: I1127 07:38:33.574809 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2r7r8_456beb89-e656-49ac-b6d2-ad92029ae549/speaker/0.log" Nov 27 07:38:40 crc kubenswrapper[4706]: I1127 07:38:40.389934 4706 scope.go:117] "RemoveContainer" containerID="31d6c1ade58f373e3a672e680187bff8f7098ad5d1c6f2cc60431e79fb4927ca" Nov 27 07:38:45 crc kubenswrapper[4706]: I1127 07:38:45.353488 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_2482c5b8-e170-4a54-9505-d422bc31d390/openstackclient/0.log" Nov 27 07:38:55 crc kubenswrapper[4706]: I1127 07:38:55.782616 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/util/0.log" Nov 27 07:38:55 crc kubenswrapper[4706]: I1127 07:38:55.970510 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/pull/0.log" Nov 27 07:38:55 crc kubenswrapper[4706]: I1127 07:38:55.993865 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/util/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.055488 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/pull/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.202728 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/extract/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.205967 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/util/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.211283 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83dxspv_19aee991-c9f9-4d97-9af1-3162462f7d1c/pull/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.431734 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-utilities/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.576635 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-utilities/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.600316 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-content/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.631180 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-content/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.792391 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-utilities/0.log" Nov 27 07:38:56 crc kubenswrapper[4706]: I1127 07:38:56.814005 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/extract-content/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.002488 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qd9pp_34610560-6b02-4f75-b595-2f3f0bef67e3/extract-utilities/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.141959 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qd9pp_34610560-6b02-4f75-b595-2f3f0bef67e3/extract-utilities/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.188061 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qd9pp_34610560-6b02-4f75-b595-2f3f0bef67e3/extract-content/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.212804 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qd9pp_34610560-6b02-4f75-b595-2f3f0bef67e3/extract-content/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.340810 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-h4vk6_547944fe-57e9-4370-90a3-9c60f4d3d03b/registry-server/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.414733 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qd9pp_34610560-6b02-4f75-b595-2f3f0bef67e3/extract-utilities/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.433896 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qd9pp_34610560-6b02-4f75-b595-2f3f0bef67e3/extract-content/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.515431 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qd9pp_34610560-6b02-4f75-b595-2f3f0bef67e3/registry-server/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.607355 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fzsrg_bd7353c5-075d-4f1d-b18e-2da68cc06652/marketplace-operator/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.723874 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-utilities/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.869772 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-content/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.875633 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-utilities/0.log" Nov 27 07:38:57 crc kubenswrapper[4706]: I1127 07:38:57.875652 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-content/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.069401 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-utilities/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.086799 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/extract-content/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.164397 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k7j8m_6c118696-a09a-405f-89e4-aedbe69487f1/registry-server/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.249716 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-utilities/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.380705 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-content/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.403086 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-utilities/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.413545 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-content/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.589444 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-content/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.596826 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/extract-utilities/0.log" Nov 27 07:38:58 crc kubenswrapper[4706]: I1127 07:38:58.878675 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n274h_3b239571-088a-4b5f-ad24-39500fad1dea/registry-server/0.log" Nov 27 07:39:26 crc kubenswrapper[4706]: E1127 07:39:26.692756 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:39:26 crc kubenswrapper[4706]: E1127 07:39:26.693396 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:41:28.693381991 +0000 UTC m=+1972.582972801 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:39:26 crc kubenswrapper[4706]: E1127 07:39:26.693017 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:39:26 crc kubenswrapper[4706]: E1127 07:39:26.693580 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:41:28.693515026 +0000 UTC m=+1972.583105876 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:40:00 crc kubenswrapper[4706]: I1127 07:40:00.512792 4706 generic.go:334] "Generic (PLEG): container finished" podID="0835baf9-d98f-44a6-af44-32e172968f2d" containerID="86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc" exitCode=0 Nov 27 07:40:00 crc kubenswrapper[4706]: I1127 07:40:00.513012 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" event={"ID":"0835baf9-d98f-44a6-af44-32e172968f2d","Type":"ContainerDied","Data":"86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc"} Nov 27 07:40:00 crc kubenswrapper[4706]: I1127 07:40:00.514173 4706 scope.go:117] "RemoveContainer" containerID="86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc" Nov 27 07:40:01 crc kubenswrapper[4706]: I1127 07:40:01.502988 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mg2kk_must-gather-hwfz2_0835baf9-d98f-44a6-af44-32e172968f2d/gather/0.log" Nov 27 07:40:10 crc kubenswrapper[4706]: I1127 07:40:10.753904 4706 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mg2kk/must-gather-hwfz2"] Nov 27 07:40:10 crc kubenswrapper[4706]: I1127 07:40:10.754768 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" podUID="0835baf9-d98f-44a6-af44-32e172968f2d" containerName="copy" containerID="cri-o://0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156" gracePeriod=2 Nov 27 07:40:10 crc kubenswrapper[4706]: I1127 07:40:10.758111 4706 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mg2kk/must-gather-hwfz2"] Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.116291 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mg2kk_must-gather-hwfz2_0835baf9-d98f-44a6-af44-32e172968f2d/copy/0.log" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.117164 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.130735 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0835baf9-d98f-44a6-af44-32e172968f2d-must-gather-output\") pod \"0835baf9-d98f-44a6-af44-32e172968f2d\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.130859 4706 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkwnt\" (UniqueName: \"kubernetes.io/projected/0835baf9-d98f-44a6-af44-32e172968f2d-kube-api-access-nkwnt\") pod \"0835baf9-d98f-44a6-af44-32e172968f2d\" (UID: \"0835baf9-d98f-44a6-af44-32e172968f2d\") " Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.136154 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0835baf9-d98f-44a6-af44-32e172968f2d-kube-api-access-nkwnt" (OuterVolumeSpecName: "kube-api-access-nkwnt") pod "0835baf9-d98f-44a6-af44-32e172968f2d" (UID: "0835baf9-d98f-44a6-af44-32e172968f2d"). InnerVolumeSpecName "kube-api-access-nkwnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.192026 4706 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0835baf9-d98f-44a6-af44-32e172968f2d-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0835baf9-d98f-44a6-af44-32e172968f2d" (UID: "0835baf9-d98f-44a6-af44-32e172968f2d"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.232485 4706 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0835baf9-d98f-44a6-af44-32e172968f2d-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.232524 4706 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkwnt\" (UniqueName: \"kubernetes.io/projected/0835baf9-d98f-44a6-af44-32e172968f2d-kube-api-access-nkwnt\") on node \"crc\" DevicePath \"\"" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.594601 4706 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mg2kk_must-gather-hwfz2_0835baf9-d98f-44a6-af44-32e172968f2d/copy/0.log" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.595035 4706 generic.go:334] "Generic (PLEG): container finished" podID="0835baf9-d98f-44a6-af44-32e172968f2d" containerID="0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156" exitCode=143 Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.595135 4706 scope.go:117] "RemoveContainer" containerID="0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.595178 4706 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mg2kk/must-gather-hwfz2" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.610639 4706 scope.go:117] "RemoveContainer" containerID="86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.657910 4706 scope.go:117] "RemoveContainer" containerID="0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156" Nov 27 07:40:11 crc kubenswrapper[4706]: E1127 07:40:11.658664 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156\": container with ID starting with 0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156 not found: ID does not exist" containerID="0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.658698 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156"} err="failed to get container status \"0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156\": rpc error: code = NotFound desc = could not find container \"0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156\": container with ID starting with 0f63d661da7c01e3993250df146665da0bc61bfa7e1ed811ad0f104d48603156 not found: ID does not exist" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.658717 4706 scope.go:117] "RemoveContainer" containerID="86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc" Nov 27 07:40:11 crc kubenswrapper[4706]: E1127 07:40:11.659208 4706 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc\": container with ID starting with 86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc not found: ID does not exist" containerID="86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc" Nov 27 07:40:11 crc kubenswrapper[4706]: I1127 07:40:11.659357 4706 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc"} err="failed to get container status \"86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc\": rpc error: code = NotFound desc = could not find container \"86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc\": container with ID starting with 86db08ef8c99c83b8bd5b7e475a6e59ed0d118280c66ce5377ebcba72b86cdfc not found: ID does not exist" Nov 27 07:40:12 crc kubenswrapper[4706]: I1127 07:40:12.783630 4706 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0835baf9-d98f-44a6-af44-32e172968f2d" path="/var/lib/kubelet/pods/0835baf9-d98f-44a6-af44-32e172968f2d/volumes" Nov 27 07:40:15 crc kubenswrapper[4706]: I1127 07:40:15.177876 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:40:15 crc kubenswrapper[4706]: I1127 07:40:15.178291 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:40:45 crc kubenswrapper[4706]: I1127 07:40:45.181512 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:40:45 crc kubenswrapper[4706]: I1127 07:40:45.182296 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:41:15 crc kubenswrapper[4706]: I1127 07:41:15.178284 4706 patch_prober.go:28] interesting pod/machine-config-daemon-c44hv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 07:41:15 crc kubenswrapper[4706]: I1127 07:41:15.178853 4706 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 07:41:15 crc kubenswrapper[4706]: I1127 07:41:15.178902 4706 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" Nov 27 07:41:15 crc kubenswrapper[4706]: I1127 07:41:15.179791 4706 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df4cd0396271a310394a11f3f8671785c69118c72eb0d461caf52dc2d37d6433"} pod="openshift-machine-config-operator/machine-config-daemon-c44hv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 07:41:15 crc kubenswrapper[4706]: I1127 07:41:15.179884 4706 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" podUID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerName="machine-config-daemon" containerID="cri-o://df4cd0396271a310394a11f3f8671785c69118c72eb0d461caf52dc2d37d6433" gracePeriod=600 Nov 27 07:41:16 crc kubenswrapper[4706]: I1127 07:41:16.017727 4706 generic.go:334] "Generic (PLEG): container finished" podID="f89d12ee-9aae-48da-a7a6-130c7186f506" containerID="df4cd0396271a310394a11f3f8671785c69118c72eb0d461caf52dc2d37d6433" exitCode=0 Nov 27 07:41:16 crc kubenswrapper[4706]: I1127 07:41:16.017822 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerDied","Data":"df4cd0396271a310394a11f3f8671785c69118c72eb0d461caf52dc2d37d6433"} Nov 27 07:41:16 crc kubenswrapper[4706]: I1127 07:41:16.018338 4706 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c44hv" event={"ID":"f89d12ee-9aae-48da-a7a6-130c7186f506","Type":"ContainerStarted","Data":"496b447e904849f8f7119119459e98ac9c2a3cbdb385d3914be2a613461aaa9e"} Nov 27 07:41:16 crc kubenswrapper[4706]: I1127 07:41:16.018358 4706 scope.go:117] "RemoveContainer" containerID="5105a6dc009ff023c9047965ce8d53e451bd1d620ee1baf4afd0a2a311cf5900" Nov 27 07:41:28 crc kubenswrapper[4706]: E1127 07:41:28.788830 4706 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 27 07:41:28 crc kubenswrapper[4706]: E1127 07:41:28.790388 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:43:30.790351124 +0000 UTC m=+2094.679941944 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config-secret") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : secret "openstack-config-secret" not found Nov 27 07:41:28 crc kubenswrapper[4706]: E1127 07:41:28.788830 4706 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 27 07:41:28 crc kubenswrapper[4706]: E1127 07:41:28.790467 4706 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config podName:2482c5b8-e170-4a54-9505-d422bc31d390 nodeName:}" failed. No retries permitted until 2025-11-27 07:43:30.790452987 +0000 UTC m=+2094.680043907 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/2482c5b8-e170-4a54-9505-d422bc31d390-openstack-config") pod "openstackclient" (UID: "2482c5b8-e170-4a54-9505-d422bc31d390") : configmap "openstack-config" not found Nov 27 07:41:40 crc kubenswrapper[4706]: I1127 07:41:40.497959 4706 scope.go:117] "RemoveContainer" containerID="8a39e365a4f5cacee0b627ef0fe67e56d2c047d5157a385785bee18728a1b1f0" Nov 27 07:41:40 crc kubenswrapper[4706]: I1127 07:41:40.527514 4706 scope.go:117] "RemoveContainer" containerID="5a1c939e2377fae10a4905456040228ea9450dedcc8eba4f2aa0850bed1e3c57" Nov 27 07:41:40 crc kubenswrapper[4706]: I1127 07:41:40.555885 4706 scope.go:117] "RemoveContainer" containerID="b80e0852a7fff52d1cd86a93d7d7e576c7a3edb068c90b60a543a012aae838ac" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112000363024434 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112000364017352 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111774112016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111774112015455 5ustar corecore